[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/osism/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T00:02:09.132306Z",
                    "start": "2026-01-13T00:00:21.367493Z"
                },
                "id": "0242ac1f-6606-b1ff-a440-000000000002",
                "name": "Base pre"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:21.443037Z",
                            "start": "2026-01-13T00:00:21.406352Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000000c",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "ca7/osism/ca76e7f80cbb4cdda68907de4afef11c"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000087",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:21.604765Z",
                            "start": "2026-01-13T00:00:21.541179Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000089",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.14\nJob: testbed-deploy-current-in-a-nutshell-with-tempest-ubuntu-24.04\nPipeline: periodic-midnight\nExecutor: 521e9411259a\nTriggered by: https://github.com/osism/testbed\nEvent ID: 266de1de9ea347ba87510853df01b929\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:21.718117Z",
                            "start": "2026-01-13T00:00:21.632153Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000000d",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: orchestrator\nHostname: zuul-static-regiocloud-infra-1\nUsername: zuul-testbed05\nDistro: Debian 12.13\nProvider: static-testbed\nRegion: \nLabel: testbed-orchestrator\nProduct Name: OpenStack Nova\nInterface IP: 81.163.193.140\n",
                                    "zj_item": "orchestrator"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:22.077450Z",
                            "start": "2026-01-13T00:00:21.725726Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000000e",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:23.931655Z",
                            "start": "2026-01-13T00:00:22.104865Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000013",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "c8280919550a8f564ac94f60a18cc00badb66456",
                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "c8280919550a8f564ac94f60a18cc00badb66456",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262424.042463-34-101258610717258/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "ebdfd0dceeb6f7194517586d005a94b2",
                            "mode": "0644",
                            "owner": "root",
                            "size": 8206,
                            "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262424.042463-34-101258610717258/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:28.081270Z",
                            "start": "2026-01-13T00:00:23.940390Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000014",
                        "name": "Copy ansible inventory to logs dir"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:29.246292Z",
                            "start": "2026-01-13T00:00:28.133287Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:29.300110Z",
                            "start": "2026-01-13T00:00:29.252270Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:01.675632",
                            "end": "2026-01-13 00:00:32.450037",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-b1ff-a440-0000000000aa-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:00:30.774405",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa.pub\nThe key fingerprint is:\nSHA256:5YhT0SjeywBjjwG3Wb4O5DgTueUtZW0cZGAjkPMaVdQ zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|  o+ooB+=o       |\n|  oo*B.=Eo.      |\n|  o==B+o= .      |\n|  .Oo++=.+       |\n|  =o= =oS..      |\n|  .o + .o        |\n|      .          |\n|                 |\n|                 |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:5YhT0SjeywBjjwG3Wb4O5DgTueUtZW0cZGAjkPMaVdQ zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|  o+ooB+=o       |",
                                "|  oo*B.=Eo.      |",
                                "|  o==B+o= .      |",
                                "|  .Oo++=.+       |",
                                "|  =o= =oS..      |",
                                "|  .o + .o        |",
                                "|      .          |",
                                "|                 |",
                                "|                 |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac1f-6606-b1ff-a440-0000000000aa-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:32.578811Z",
                            "start": "2026-01-13T00:00:29.385176Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000aa",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:32.670675Z",
                            "start": "2026-01-13T00:00:32.585466Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000ab",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:32.805765Z",
                            "start": "2026-01-13T00:00:32.732058Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000ca",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCh/WKRmWvvHxojusJHTg1MCf5FkUfeP7ZQuitGz7rJpJE/3SzpC9mi7uddKyfgOEjV6y/xk9et5urJTN10nh4RxJuAesMu9VcoNOmMA/7/Z1VhWYKg2RcCQ/KTaCgd/MwbuwLywAxeV7JIWHfP65f0s9u0BzLbrET2e1T74XrnUA2qoMsa7u9QWJV8nZsQfiqummY3Az+hgPQ/+R7XXK9fam6CL2SidqB0v7dfwDnZZjLsoBfrGpvDKyEEm7DkSyhvqaasBPHSWTn6qe0UV6yXaMKTUnkrxKcBmpjgwR+o0ipHGVo0Ocnpm+aPm8Tq4rIMLLykDkzms6HS9rXvS8P54lawn2Ou2yWlTOjkNDJlJJyeu8pE1BxvMAq5CNw1Xt1VfmM1Lawbm1AKa7kL29LNyzvNHPZ68/kS8uL7+NKEjV88b7Y6yuX0AQR5mt6unQUntEGWZwLW2OlPoxHKQcMXr3yT/CDCaCX6K48Tzttncyl9yThpKlq2Wf8mIQbGBG8= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed05/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul-testbed05",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCh/WKRmWvvHxojusJHTg1MCf5FkUfeP7ZQuitGz7rJpJE/3SzpC9mi7uddKyfgOEjV6y/xk9et5urJTN10nh4RxJuAesMu9VcoNOmMA/7/Z1VhWYKg2RcCQ/KTaCgd/MwbuwLywAxeV7JIWHfP65f0s9u0BzLbrET2e1T74XrnUA2qoMsa7u9QWJV8nZsQfiqummY3Az+hgPQ/+R7XXK9fam6CL2SidqB0v7dfwDnZZjLsoBfrGpvDKyEEm7DkSyhvqaasBPHSWTn6qe0UV6yXaMKTUnkrxKcBmpjgwR+o0ipHGVo0Ocnpm+aPm8Tq4rIMLLykDkzms6HS9rXvS8P54lawn2Ou2yWlTOjkNDJlJJyeu8pE1BxvMAq5CNw1Xt1VfmM1Lawbm1AKa7kL29LNyzvNHPZ68/kS8uL7+NKEjV88b7Y6yuX0AQR5mt6unQUntEGWZwLW2OlPoxHKQcMXr3yT/CDCaCX6K48Tzttncyl9yThpKlq2Wf8mIQbGBG8= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed05/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul-testbed05",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:34.106722Z",
                            "start": "2026-01-13T00:00:32.820121Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000cb",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed05/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed05/.ssh"
                                }
                            },
                            "gid": 1006,
                            "group": "zuul-testbed05",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed05/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul-testbed05",
                            "path": "/home/zuul-testbed05/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1005
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:34.441091Z",
                            "start": "2026-01-13T00:00:34.133796Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000cc",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed05/.ssh/id_rsa",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa",
                                "force": false,
                                "mode": 384,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa",
                                    "force": false,
                                    "mode": 384,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa"
                                },
                                "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa"
                            },
                            "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:34.971194Z",
                            "start": "2026-01-13T00:00:34.449376Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000cd",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed05/.ssh/id_rsa.pub",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa.pub",
                                "force": false,
                                "mode": 420,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa.pub",
                                    "force": false,
                                    "mode": 420,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa.pub"
                                },
                                "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa.pub"
                            },
                            "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa.pub"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:35.547664Z",
                            "start": "2026-01-13T00:00:34.998135Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000ce",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:35.588058Z",
                            "start": "2026-01-13T00:00:35.553732Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000ac",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/var/ssh/nodepool"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-0000000000b1",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:37.458340Z",
                            "start": "2026-01-13T00:00:35.595482Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000b3",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa"
                            ],
                            "delta": "0:00:00.034179",
                            "end": "2026-01-13 00:00:38.679752",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-b1ff-a440-0000000000b9-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:00:38.645573",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/ca76e7f80cbb4cdda68907de4afef11c_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-b1ff-a440-0000000000b9-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:38.797509Z",
                            "start": "2026-01-13T00:00:37.483526Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000b9",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:39.715209Z",
                            "start": "2026-01-13T00:00:38.804302Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000ba",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:39.798075Z",
                            "start": "2026-01-13T00:00:39.728101Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-0000000000bb",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001b",
                        "name": "start-zuul-console",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/start-zuul-console"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:40.668198Z",
                            "start": "2026-01-13T00:00:40.016059Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_info_dir": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:40.723855Z",
                            "start": "2026-01-13T00:00:40.691657Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000021",
                        "name": "Define zuul_info_dir fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:41.518611Z",
                            "start": "2026-01-13T00:00:40.734974Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000022",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "setup",
                            "ansible_facts": {
                                "ansible_all_ipv4_addresses": [
                                    "10.128.0.21"
                                ],
                                "ansible_all_ipv6_addresses": [
                                    "2a13:1a81:8000:2::33d",
                                    "fe80::f816:3eff:febf:60f0"
                                ],
                                "ansible_apparmor": {
                                    "status": "enabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "04/01/2014",
                                "ansible_bios_vendor": "SeaBIOS",
                                "ansible_bios_version": "1.15.0-1",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "QEMU",
                                "ansible_chassis_version": "pc-i440fx-6.2",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": "ttyS0,115200",
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_date_time": {
                                    "date": "2026-01-13",
                                    "day": "13",
                                    "epoch": "1768262442",
                                    "epoch_int": "1768262442",
                                    "hour": "00",
                                    "iso8601": "2026-01-13T00:00:42Z",
                                    "iso8601_basic": "20260113T000042945430",
                                    "iso8601_basic_short": "20260113T000042",
                                    "iso8601_micro": "2026-01-13T00:00:42.945430Z",
                                    "minute": "00",
                                    "month": "01",
                                    "second": "42",
                                    "time": "00:00:42",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Tuesday",
                                    "weekday_number": "2",
                                    "weeknumber": "02",
                                    "year": "2026"
                                },
                                "ansible_default_ipv4": {
                                    "address": "10.128.0.21",
                                    "alias": "ens3",
                                    "broadcast": "",
                                    "gateway": "10.128.0.17",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "netmask": "255.255.255.240",
                                    "network": "10.128.0.16",
                                    "prefix": "28",
                                    "type": "ether"
                                },
                                "ansible_default_ipv6": {
                                    "address": "2a13:1a81:8000:2::33d",
                                    "gateway": "fe80::f816:3eff:fe98:d9f",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "prefix": "128",
                                    "scope": "global",
                                    "type": "ether"
                                },
                                "ansible_device_links": {
                                    "ids": {
                                        "sda": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                        ],
                                        "sda1": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                        ],
                                        "sda14": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                        ],
                                        "sda15": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                        ]
                                    },
                                    "labels": {},
                                    "masters": {},
                                    "uuids": {
                                        "sda1": [
                                            "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                        ],
                                        "sda15": [
                                            "1E24-44C9"
                                        ]
                                    }
                                },
                                "ansible_devices": {
                                    "loop0": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop1": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop2": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop3": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop4": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop5": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop6": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop7": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "sda": {
                                        "holders": [],
                                        "host": "SCSI storage controller: Red Hat, Inc. Virtio SCSI",
                                        "links": {
                                            "ids": [
                                                "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                            ],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": "QEMU HARDDISK",
                                        "partitions": {
                                            "sda1": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                                    ]
                                                },
                                                "sectors": "209453023",
                                                "sectorsize": 512,
                                                "size": "99.87 GB",
                                                "start": "262144",
                                                "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                            },
                                            "sda14": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": []
                                                },
                                                "sectors": "6144",
                                                "sectorsize": 512,
                                                "size": "3.00 MB",
                                                "start": "2048",
                                                "uuid": null
                                            },
                                            "sda15": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "1E24-44C9"
                                                    ]
                                                },
                                                "sectors": "253952",
                                                "sectorsize": 512,
                                                "size": "124.00 MB",
                                                "start": "8192",
                                                "uuid": "1E24-44C9"
                                            }
                                        },
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "209715200",
                                        "sectorsize": "512",
                                        "size": "100.00 GB",
                                        "support_discard": "4096",
                                        "vendor": "QEMU",
                                        "virtual": 1
                                    }
                                },
                                "ansible_distribution": "Debian",
                                "ansible_distribution_file_parsed": true,
                                "ansible_distribution_file_path": "/etc/os-release",
                                "ansible_distribution_file_variety": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_minor_version": "13",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12.13",
                                "ansible_dns": {
                                    "nameservers": [
                                        "81.163.194.9",
                                        "81.163.194.10",
                                        "2a13:1a81::11:53",
                                        "2a13:1a81::12:53"
                                    ],
                                    "search": [
                                        "."
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 1006,
                                "ansible_effective_user_id": 1005,
                                "ansible_ens3": {
                                    "active": true,
                                    "device": "ens3",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "off [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "off [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "on",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "on [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "off [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "off [fixed]",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "on [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "off [fixed]",
                                        "tx_nocache_copy": "off",
                                        "tx_scatter_gather": "on",
                                        "tx_scatter_gather_fraglist": "off [fixed]",
                                        "tx_sctp_segmentation": "off [fixed]",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "off",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "off [fixed]",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "off [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "10.128.0.21",
                                        "broadcast": "",
                                        "netmask": "255.255.255.240",
                                        "network": "10.128.0.16",
                                        "prefix": "28"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "2a13:1a81:8000:2::33d",
                                            "prefix": "128",
                                            "scope": "global"
                                        },
                                        {
                                            "address": "fe80::f816:3eff:febf:60f0",
                                            "prefix": "64",
                                            "scope": "link"
                                        }
                                    ],
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "module": "virtio_net",
                                    "mtu": 1500,
                                    "pciid": "virtio1",
                                    "promisc": false,
                                    "speed": -1,
                                    "timestamping": [],
                                    "type": "ether"
                                },
                                "ansible_env": {
                                    "HOME": "/home/zuul-testbed05",
                                    "LANG": "C.UTF-8",
                                    "LOGNAME": "zuul-testbed05",
                                    "MOTD_SHOWN": "pam",
                                    "PATH": "/usr/local/bin:/usr/bin:/bin:/usr/games",
                                    "PWD": "/home/zuul-testbed05",
                                    "SHELL": "/bin/sh",
                                    "SSH_CLIENT": "81.163.193.69 55836 22",
                                    "SSH_CONNECTION": "81.163.193.69 55836 10.128.0.21 22",
                                    "USER": "zuul-testbed05",
                                    "XDG_RUNTIME_DIR": "/run/user/1005",
                                    "XDG_SESSION_CLASS": "user",
                                    "XDG_SESSION_ID": "17877",
                                    "XDG_SESSION_TYPE": "tty"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "Other",
                                "ansible_fqdn": "zuul-static-regiocloud-infra-1",
                                "ansible_hostname": "zuul-static-regiocloud-infra-1",
                                "ansible_hostnqn": "",
                                "ansible_interfaces": [
                                    "ens3",
                                    "lo"
                                ],
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "6.1.0-23-cloud-amd64",
                                "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15)",
                                "ansible_lo": {
                                    "active": true,
                                    "device": "lo",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "on [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "on [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "off [fixed]",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "off [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on [fixed]",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "on [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "on",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "off [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "on [fixed]",
                                        "tx_nocache_copy": "off [fixed]",
                                        "tx_scatter_gather": "on [fixed]",
                                        "tx_scatter_gather_fraglist": "on [fixed]",
                                        "tx_sctp_segmentation": "on",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "on",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "on",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "on [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "127.0.0.1",
                                        "broadcast": "",
                                        "netmask": "255.0.0.0",
                                        "network": "127.0.0.0",
                                        "prefix": "8"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "::1",
                                            "prefix": "128",
                                            "scope": "host"
                                        }
                                    ],
                                    "mtu": 65536,
                                    "promisc": false,
                                    "timestamping": [],
                                    "type": "loopback"
                                },
                                "ansible_loadavg": {
                                    "15m": 0.0048828125,
                                    "1m": 0.0732421875,
                                    "5m": 0.01611328125
                                },
                                "ansible_local": {},
                                "ansible_locally_reachable_ips": {
                                    "ipv4": [
                                        "10.128.0.21",
                                        "127.0.0.0/8",
                                        "127.0.0.1"
                                    ],
                                    "ipv6": [
                                        "::1",
                                        "2a13:1a81:8000:2::33d",
                                        "fe80::f816:3eff:febf:60f0"
                                    ]
                                },
                                "ansible_lsb": {
                                    "codename": "bookworm",
                                    "description": "Debian GNU/Linux 12 (bookworm)",
                                    "id": "Debian",
                                    "major_release": "12",
                                    "release": "12"
                                },
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_machine_id": "28f0fb0ddd3c48c38b3aa31603cb444d",
                                "ansible_memfree_mb": 503,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 7286,
                                        "used": 666
                                    },
                                    "real": {
                                        "free": 503,
                                        "total": 7952,
                                        "used": 7449
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 7952,
                                "ansible_mounts": [
                                    {
                                        "block_available": 17831098,
                                        "block_size": 4096,
                                        "block_total": 25759134,
                                        "block_used": 7928036,
                                        "device": "/dev/sda1",
                                        "fstype": "ext4",
                                        "inode_available": 5920387,
                                        "inode_total": 6545408,
                                        "inode_used": 625021,
                                        "mount": "/",
                                        "options": "rw,relatime,discard,errors=remount-ro",
                                        "size_available": 73036177408,
                                        "size_total": 105509412864,
                                        "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                    },
                                    {
                                        "block_available": 57419,
                                        "block_size": 2048,
                                        "block_total": 63339,
                                        "block_used": 5920,
                                        "device": "/dev/sda15",
                                        "fstype": "vfat",
                                        "inode_available": 0,
                                        "inode_total": 0,
                                        "inode_used": 0,
                                        "mount": "/boot/efi",
                                        "options": "rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=ascii,shortname=mixed,utf8,errors=remount-ro",
                                        "size_available": 117594112,
                                        "size_total": 129718272,
                                        "uuid": "1E24-44C9"
                                    }
                                ],
                                "ansible_nodename": "zuul-static-regiocloud-infra-1",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": [
                                        "tty0",
                                        "ttyS0,115200"
                                    ],
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_processor": [
                                    "0",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "1",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "2",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "3",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 4,
                                "ansible_processor_nproc": 4,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 4,
                                "ansible_product_name": "OpenStack Nova",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "26.2.1",
                                "ansible_python": {
                                    "executable": "/usr/bin/python3",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 2,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        2,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.2",
                                "ansible_real_group_id": 1006,
                                "ansible_real_user_id": 1005,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "systemd",
                                "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAOVugYBPOvEOPI9DA30PvhAEr88mJeaFmKN4pkfnBHV/1A1OlDKtX7lKWCBfEUilcd+tQFVKXm4KY+K9q4J5x5tpd7cL7LRDCFKN+py3+jc6U5J9EZ0o5cVG84z0c3XLEL6cCljG9Etx/0pXjjELRh1SGuUZweEx0143OSiKs/9DAAAAFQCT7R5MDTPEdFXQcZxA/fPRmtWFswAAAIBp8Iz2Njfcxfnh/mdxMztzkdNXulDt679+ghvl5588eCIDPa9xMtojivk53JaV6V4R+wsBN+MaKufzMOoVxuDkZcmw940H/1kcr3yOsJcZIkU1GRkJiNnSEIWJ6dVaXMh+wwbjw+8xNCA3+n3fARqUK906ICT5RsRrBe7W2K7dIAAAAIAL9qFr0vTidcXXhyWKN8I2iBsHSKfb6cjbseShw3ZeZE+kJ3IXz3z9oe0OwKp52GhvJrOXEkxST9YbtutbsVz5iS7lDpisvIWfFtvRRYaOtyvE8b4PWpHrk4s5tssqGDQ7y7zSqmRrj1sf5w0pwokPGmoSnPWUi0Vp0yhlysI/1w==",
                                "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss",
                                "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAuyO5R2SR4XT/Y65sItV7OQrUJXmGJ+//L9L/VugDohOWInFrkaZ6yFJMccVS7fmApekshaCGYnkIVdXV72hzg=",
                                "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256",
                                "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAILxLVCAcH99lPquKCT3PwHsvi+569sV28I73LxchEjeg",
                                "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519",
                                "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ3YX5Qhzq+HT8jTGGcX7rLsF3GCwObHFepfMmBZuD+gTfvdY5sYFSoXiT70rLPBjqETBTS0SQwmo8sd5AJI1vdgIqI5Sry8p52ng8tDg6C5CONt+xAaYOmE+ZDD/6naNX56b/gmgqBRQqPQHPmuptNxTlimdAcpoi+p9/UFBLTObMYmNL+5DTqw1+XECJ5j5kj5Ez40NT5w/1XnelEMcdw06WJZrzNEFt4Qpr2Zf+ybh0s+Xcn9u7d/gCbwXIPJvJnT0iSkiOzFvXJgOzwpvPtoi/5cV7g1LmgWsMLadT6gHof2poQaofC+1Bi0KiB2ANARMumvVBULx03qKvsqulsnlQAix2ffMraa3aMdUzK5aZo3hBx+H1uyEznUlyG9TUT5zhLIhNObetbpJgv6OJHiknBW6v/80FGD9t1pDiAwwBc/tJ8TyKpwyAdGMiEBIIVxTzx8QfPuApPJBJZXhMo2FBwRAEDEuuHbZu0znjhKKz4zlFT8TDrW+W7iThTvc=",
                                "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": [
                                    ""
                                ],
                                "ansible_system_capabilities_enforced": "True",
                                "ansible_system_vendor": "OpenStack Foundation",
                                "ansible_uptime_seconds": 43592418,
                                "ansible_user_dir": "/home/zuul-testbed05",
                                "ansible_user_gecos": "",
                                "ansible_user_gid": 1006,
                                "ansible_user_id": "zuul-testbed05",
                                "ansible_user_shell": "/bin/sh",
                                "ansible_user_uid": 1005,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "guest",
                                "ansible_virtualization_tech_guest": [
                                    "openstack"
                                ],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "openstack",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "fact_path": "/etc/ansible/facts.d",
                                    "filter": [],
                                    "gather_subset": [
                                        "all"
                                    ],
                                    "gather_timeout": 10
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:43.321572Z",
                            "start": "2026-01-13T00:00:41.524470Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000023",
                        "name": "Collect information about the host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "validate_host_hostname_sanitized": "orchestrator"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:43.488141Z",
                            "start": "2026-01-13T00:00:43.329003Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000024",
                        "name": "Sanitize hostname"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "1b5e10ca2c771c7fdf9d3785ef09806f14bd7506",
                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info/host-info.orchestrator.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "host-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "1b5e10ca2c771c7fdf9d3785ef09806f14bd7506",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info/host-info.orchestrator.yaml",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262443.6055417-238-122840241916615/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "963d3afb5eafb90ad235b554d35e1bdc",
                            "mode": "0644",
                            "owner": "root",
                            "size": 19731,
                            "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262443.6055417-238-122840241916615/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:45.610685Z",
                            "start": "2026-01-13T00:00:43.492577Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000025",
                        "name": "Write out all ansible variables/facts known for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_zuul_nolog_return": true,
                            "action": "zuul_debug_info",
                            "changed": false,
                            "df_i": "Filesystem      Inodes  IUsed   IFree IUse% Mounted on\nudev           1015018    312 1014706    1% /dev\ntmpfs          1017896    541 1017355    1% /run\n/dev/sda1      6545408 625023 5920385   10% /\ntmpfs          1017896      1 1017895    1% /dev/shm\ntmpfs          1017896      2 1017894    1% /run/lock\n/dev/sda15           0      0       0     - /boot/efi\ntmpfs           203579     17  203562    1% /run/user/1001\ntmpfs           203579     17  203562    1% /run/user/1004\ntmpfs           203579     17  203562    1% /run/user/1005\ntmpfs           203579     17  203562    1% /run/user/1006\n",
                            "df_m": "Filesystem     1M-blocks  Used Available Use% Mounted on\nudev                3965     0      3965   0% /dev\ntmpfs                796     1       795   1% /run\n/dev/sda1         100622 26843     69653  28% /\ntmpfs               3977     0      3977   0% /dev/shm\ntmpfs                  5     0         5   0% /run/lock\n/dev/sda15           124    12       113  10% /boot/efi\ntmpfs                796     0       796   0% /run/user/1001\ntmpfs                796     0       796   0% /run/user/1004\ntmpfs                796     0       796   0% /run/user/1005\ntmpfs                796     0       796   0% /run/user/1006\n",
                            "image_manifest_files": [],
                            "invocation": {
                                "module_args": {
                                    "image_manifest": null,
                                    "image_manifest_files": [
                                        "/etc/dib-builddate.txt",
                                        "/etc/image-hostname.txt"
                                    ],
                                    "ipv4_route_required": false,
                                    "ipv6_route_required": false,
                                    "traceroute_host": null
                                }
                            },
                            "network_interfaces": "1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n    inet 127.0.0.1/8 scope host lo\n       valid_lft forever preferred_lft forever\n    inet6 ::1/128 scope host noprefixroute \n       valid_lft forever preferred_lft forever\n2: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000\n    link/ether fa:16:3e:bf:60:f0 brd ff:ff:ff:ff:ff:ff\n    altname enp0s3\n    inet 10.128.0.21/28 metric 100 brd 10.128.0.31 scope global dynamic ens3\n       valid_lft 24822sec preferred_lft 24822sec\n    inet6 2a13:1a81:8000:2::33d/128 scope global noprefixroute \n       valid_lft forever preferred_lft forever\n    inet6 fe80::f816:3eff:febf:60f0/64 scope link \n       valid_lft forever preferred_lft forever\n",
                            "network_neighbors": "10.128.0.17 dev ens3 lladdr fa:16:3e:e3:0c:52 REACHABLE \nfe80::f816:3eff:fe98:d9f dev ens3 lladdr fa:16:3e:98:0d:9f router STALE \n",
                            "network_routing_v4": "default via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n10.128.0.16/28 dev ens3 proto kernel scope link src 10.128.0.21 metric 100 \n10.128.0.17 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n10.128.0.18 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n81.163.194.9 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n81.163.194.10 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n169.254.169.254 via 10.128.0.18 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n",
                            "network_routing_v6": "2a13:1a81:8000:2::300/120 dev ens3 proto ra metric 100 pref medium\nfe80::/64 dev ens3 proto kernel metric 256 pref medium\ndefault via fe80::f816:3eff:fe98:d9f dev ens3 proto ra metric 100 mtu 1500 pref medium\n",
                            "proc_cpuinfo": "processor\t: 0\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 0\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 0\ninitial apicid\t: 0\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 1\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 1\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 1\ninitial apicid\t: 1\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 2\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 2\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 2\ninitial apicid\t: 2\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 3\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 3\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 3\ninitial apicid\t: 3\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\n",
                            "traceroute": null,
                            "uname": "Linux zuul-static-regiocloud-infra-1 6.1.0-23-cloud-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:46.306318Z",
                            "start": "2026-01-13T00:00:45.616162Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000027",
                        "name": "Collect information about zuul worker"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "2963fe4aeecf478c000b0d5b206500f6f638f40c",
                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info/zuul-info.orchestrator.txt",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "zuul-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "2963fe4aeecf478c000b0d5b206500f6f638f40c",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/zuul-info/zuul-info.orchestrator.txt",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262446.478094-277-222376381604849/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "b9db13a2e5e12fce81bb89c79a70699a",
                            "mode": "0644",
                            "owner": "root",
                            "size": 3369,
                            "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/.ansible/tmp/ansible-tmp-1768262446.478094-277-222376381604849/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:48.063150Z",
                            "start": "2026-01-13T00:00:46.327205Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-000000000028",
                        "name": "Write out all zuul information for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:00:48.416786Z",
                            "start": "2026-01-13T00:00:48.074915Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000002c",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --compress --delete-after --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --out-format='<<CHANGED>>%i %n%L' /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/src zuul-testbed05@81.163.193.140:.",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": false,
                                    "delete": true,
                                    "dest": "zuul-testbed05@81.163.193.140:.",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": false,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "push",
                                    "owner": false,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": true,
                                    "rsync_opts": [],
                                    "rsync_path": null,
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/src",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": ".d..t...... src/\n.d..t...... src/github.com/\n.d..t...... src/github.com/osism/\n.d..t...... src/github.com/osism/ansible-collection-commons/\n<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-commons/.flake8\n<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-commons/README.md\n<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/description\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/index\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/\n<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-services/\n<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-services/.flake8\n<f..t...... src/github.com/osism/ansible-collection-services/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-services/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-services/README.md\n<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-services/.git/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-services/.git/description\n<f..t...... src/github.com/osism/ansible-collection-services/.git/index\n<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-services/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-services/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/47/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.github/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-services/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-services/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2\n.d..t...... src/github.com/osism/terraform-base/\n<f..t...... src/github.com/osism/terraform-base/.yamllint.yml\n<f..t...... src/github.com/osism/terraform-base/.zuul.yaml\n<f..t...... src/github.com/osism/terraform-base/LICENSE\n<f..t...... src/github.com/osism/terraform-base/README.md\n.d..t...... src/github.com/osism/terraform-base/.git/\n<f..t...... src/github.com/osism/terraform-base/.git/HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/config\n<f..t...... src/github.com/osism/terraform-base/.git/config.worktree\n<f..t...... src/github.com/osism/terraform-base/.git/description\n<f..t...... src/github.com/osism/terraform-base/.git/index\n<f..t...... src/github.com/osism/terraform-base/.git/packed-refs\n.d..t...... src/github.com/osism/terraform-base/.git/branches/\n.d..t...... src/github.com/osism/terraform-base/.git/hooks/\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample\n.d..t...... src/github.com/osism/terraform-base/.git/info/\n<f..t...... src/github.com/osism/terraform-base/.git/info/exclude\n.d..t...... src/github.com/osism/terraform-base/.git/logs/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/objects/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/00/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/01/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/02/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/06/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/07/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/08/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/10/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/12/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/15/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/16/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/17/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/20/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/21/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/24/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/26/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/2c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/36/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/37/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/41/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/43/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/46/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/47/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/48/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/49/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/52/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/55/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/56/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/58/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/5d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/61/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/63/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/66/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/67/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/68/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/6d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/70/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/72/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/81/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/82/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/83/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/84/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/85/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/86/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/88/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/90/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/91/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/92/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/95/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/96/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/98/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9e/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/aa/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/af/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ba/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ce/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/da/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/db/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/de/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ea/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ee/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ef/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/fb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ff/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/info/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/pack/\n.d..t...... src/github.com/osism/terraform-base/.git/refs/\n.d..t...... src/github.com/osism/terraform-base/.github/\n<f..t...... src/github.com/osism/terraform-base/.github/renovate.json\n.d..t...... src/github.com/osism/terraform-base/playbooks/\n<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml\n.d..t...... src/github.com/osism/terraform-base/testbed-default/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/environments/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars\n.d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf\n.d..t...... src/github.com/osism/testbed/\n<f..t...... src/github.com/osism/testbed/.ansible-lint\n<f..t...... src/github.com/osism/testbed/.flake8\n<f..t...... src/github.com/osism/testbed/.gitignore\n<f..t...... src/github.com/osism/testbed/.yamllint.yml\n<f..t...... src/github.com/osism/testbed/.zuul.yaml\n<f..t...... src/github.com/osism/testbed/LICENSE\n<f..t...... src/github.com/osism/testbed/Makefile\n<f..t...... src/github.com/osism/testbed/README.md\n<f..t...... src/github.com/osism/testbed/ansible.cfg\n<f..t...... src/github.com/osism/testbed/gilt.yml\n<f..t...... src/github.com/osism/testbed/requirements.txt\n<f..t...... src/github.com/osism/testbed/requirements.yml\n<f..t...... src/github.com/osism/testbed/tox.ini\n.d..t...... src/github.com/osism/testbed/.git/\n<f..t...... src/github.com/osism/testbed/.git/HEAD\n<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/testbed/.git/config\n<f..t...... src/github.com/osism/testbed/.git/config.worktree\n<f..t...... src/github.com/osism/testbed/.git/description\n<f.st...... src/github.com/osism/testbed/.git/index\n<f..t...... src/github.com/osism/testbed/.git/packed-refs\n.d..t...... src/github.com/osism/testbed/.git/branches/\n.d..t...... src/github.com/osism/testbed/.git/hooks/\n<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample\n.d..t...... src/github.com/osism/testbed/.git/info/\n<f..t...... src/github.com/osism/testbed/.git/info/exclude\n.d..t...... src/github.com/osism/testbed/.git/logs/\n<f..t...... src/github.com/osism/testbed/.git/logs/HEAD\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/testbed/.git/objects/\n.d..t...... src/github.com/osism/testbed/.git/objects/00/\n.d..t...... src/github.com/osism/testbed/.git/objects/01/\n.d..t...... src/github.com/osism/testbed/.git/objects/02/\n.d..t...... src/github.com/osism/testbed/.git/objects/03/\n.d..t...... src/github.com/osism/testbed/.git/objects/04/\n.d..t...... src/github.com/osism/testbed/.git/objects/05/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/05/c88d3bc61d80be6a0190ba4158bc7856df0538\n.d..t...... src/github.com/osism/testbed/.git/objects/06/\n.d..t...... src/github.com/osism/testbed/.git/objects/07/\n.d..t...... src/github.com/osism/testbed/.git/objects/08/\n.d..t...... src/github.com/osism/testbed/.git/objects/09/\n.d..t...... src/github.com/osism/testbed/.git/objects/0a/\n.d..t...... src/github.com/osism/testbed/.git/objects/0b/\n.d..t...... src/github.com/osism/testbed/.git/objects/0c/\n.d..t...... src/github.com/osism/testbed/.git/objects/0d/\n.d..t...... src/github.com/osism/testbed/.git/objects/0f/\n.d..t...... src/github.com/osism/testbed/.git/objects/10/\n.d..t...... src/github.com/osism/testbed/.git/objects/11/\n.d..t...... src/github.com/osism/testbed/.git/objects/12/\n.d..t...... src/github.com/osism/testbed/.git/objects/14/\n.d..t...... src/github.com/osism/testbed/.git/objects/15/\n.d..t...... src/github.com/osism/testbed/.git/objects/16/\n.d..t...... src/github.com/osism/testbed/.git/objects/17/\n.d..t...... src/github.com/osism/testbed/.git/objects/18/\n.d..t...... src/github.com/osism/testbed/.git/objects/19/\n.d..t...... src/github.com/osism/testbed/.git/objects/1a/\n.d..t...... src/github.com/osism/testbed/.git/objects/1b/\n.d..t...... src/github.com/osism/testbed/.git/objects/1c/\n.d..t...... src/github.com/osism/testbed/.git/objects/1d/\n.d..t...... src/github.com/osism/testbed/.git/objects/20/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/20/0c611c75f445983f13b8da3e93db7d47ed81bb\n.d..t...... src/github.com/osism/testbed/.git/objects/21/\n.d..t...... src/github.com/osism/testbed/.git/objects/22/\n.d..t...... src/github.com/osism/testbed/.git/objects/23/\n.d..t...... src/github.com/osism/testbed/.git/objects/24/\n.d..t...... src/github.com/osism/testbed/.git/objects/25/\n.d..t...... src/github.com/osism/testbed/.git/objects/26/\n.d..t...... src/github.com/osism/testbed/.git/objects/27/\n.d..t...... src/github.com/osism/testbed/.git/objects/28/\n.d..t...... src/github.com/osism/testbed/.git/objects/2a/\n.d..t...... src/github.com/osism/testbed/.git/objects/2b/\n.d..t...... src/github.com/osism/testbed/.git/objects/2c/\n.d..t...... src/github.com/osism/testbed/.git/objects/2d/\n.d..t...... src/github.com/osism/testbed/.git/objects/2e/\n.d..t...... src/github.com/osism/testbed/.git/objects/2f/\n.d..t...... src/github.com/osism/testbed/.git/objects/30/\n.d..t...... src/github.com/osism/testbed/.git/objects/31/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/31/bd0867a2edfe0f397bd15bb36566a2cd7fd027\n.d..t...... src/github.com/osism/testbed/.git/objects/32/\n.d..t...... src/github.com/osism/testbed/.git/objects/33/\n.d..t...... src/github.com/osism/testbed/.git/objects/34/\n.d..t...... src/github.com/osism/testbed/.git/objects/35/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/35/7756bf2a17acc26a5f1f6c2d0a36447a7e80eb\n.d..t...... src/github.com/osism/testbed/.git/objects/36/\n.d..t...... src/github.com/osism/testbed/.git/objects/37/\n.d..t...... src/github.com/osism/testbed/.git/objects/38/\n.d..t...... src/github.com/osism/testbed/.git/objects/39/\n.d..t...... src/github.com/osism/testbed/.git/objects/3a/\n.d..t...... src/github.com/osism/testbed/.git/objects/3b/\n.d..t...... src/github.com/osism/testbed/.git/objects/3c/\n.d..t...... src/github.com/osism/testbed/.git/objects/3d/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/3d/34dfba029d7b86d4254f892fb6ac8d456e32de\n.d..t...... src/github.com/osism/testbed/.git/objects/3e/\n.d..t...... src/github.com/osism/testbed/.git/objects/3f/\n.d..t...... src/github.com/osism/testbed/.git/objects/40/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/40/46222976dc495717131ef332b39bc7929e3001\n.d..t...... src/github.com/osism/testbed/.git/objects/41/\n.d..t...... src/github.com/osism/testbed/.git/objects/42/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/42/b694dcc3f1dae4f49ac7840bff7c0ccc770b41\n.d..t...... src/github.com/osism/testbed/.git/objects/43/\n.d..t...... src/github.com/osism/testbed/.git/objects/44/\n.d..t...... src/github.com/osism/testbed/.git/objects/45/\n.d..t...... src/github.com/osism/testbed/.git/objects/46/\n.d..t...... src/github.com/osism/testbed/.git/objects/47/\n.d..t...... src/github.com/osism/testbed/.git/objects/48/\n.d..t...... src/github.com/osism/testbed/.git/objects/49/\n.d..t...... src/github.com/osism/testbed/.git/objects/4a/\n.d..t...... src/github.com/osism/testbed/.git/objects/4c/\n.d..t...... src/github.com/osism/testbed/.git/objects/4d/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/4d/515798706b72a52536f96fd021877023a4df22\n.d..t...... src/github.com/osism/testbed/.git/objects/4e/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/4e/f4687bc0e265f5692b55f6fb3b7e1ecb661918\n.d..t...... src/github.com/osism/testbed/.git/objects/4f/\n.d..t...... src/github.com/osism/testbed/.git/objects/50/\n.d..t...... src/github.com/osism/testbed/.git/objects/51/\n.d..t...... src/github.com/osism/testbed/.git/objects/52/\n.d..t...... src/github.com/osism/testbed/.git/objects/54/\n.d..t...... src/github.com/osism/testbed/.git/objects/55/\n.d..t...... src/github.com/osism/testbed/.git/objects/57/\n.d..t...... src/github.com/osism/testbed/.git/objects/58/\n.d..t...... src/github.com/osism/testbed/.git/objects/5a/\n.d..t...... src/github.com/osism/testbed/.git/objects/5b/\n.d..t...... src/github.com/osism/testbed/.git/objects/5d/\n.d..t...... src/github.com/osism/testbed/.git/objects/5e/\n.d..t...... src/github.com/osism/testbed/.git/objects/5f/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/5f/08648a61449abc0a5fd129393ef87b0b64902b\n.d..t...... src/github.com/osism/testbed/.git/objects/60/\n.d..t...... src/github.com/osism/testbed/.git/objects/61/\n.d..t...... src/github.com/osism/testbed/.git/objects/64/\n.d..t...... src/github.com/osism/testbed/.git/objects/66/\n.d..t...... src/github.com/osism/testbed/.git/objects/68/\n.d..t...... src/github.com/osism/testbed/.git/objects/69/\ncd+++++++++ src/github.com/osism/testbed/.git/objects/6a/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/6a/98d0e99dc7e391ce46704852e375bdc6ee2406\n.d..t...... src/github.com/osism/testbed/.git/objects/6b/\n.d..t...... src/github.com/osism/testbed/.git/objects/6c/\n.d..t...... src/github.com/osism/testbed/.git/objects/6e/\n.d..t...... src/github.com/osism/testbed/.git/objects/70/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/70/4ea3b380be6639d26c68e230e4acadd1b36f78\n.d..t...... src/github.com/osism/testbed/.git/objects/71/\n.d..t...... src/github.com/osism/testbed/.git/objects/72/\n.d..t...... src/github.com/osism/testbed/.git/objects/73/\n.d..t...... src/github.com/osism/testbed/.git/objects/75/\n.d..t...... src/github.com/osism/testbed/.git/objects/76/\n.d..t...... src/github.com/osism/testbed/.git/objects/77/\n.d..t...... src/github.com/osism/testbed/.git/objects/78/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/78/f8e688db2b72f61840ef15817a13b831e2e64d\n.d..t...... src/github.com/osism/testbed/.git/objects/79/\n.d..t...... src/github.com/osism/testbed/.git/objects/7a/\n.d..t...... src/github.com/osism/testbed/.git/objects/7b/\n.d..t...... src/github.com/osism/testbed/.git/objects/7c/\n.d..t...... src/github.com/osism/testbed/.git/objects/7e/\n.d..t...... src/github.com/osism/testbed/.git/objects/7f/\n.d..t...... src/github.com/osism/testbed/.git/objects/80/\n.d..t...... src/github.com/osism/testbed/.git/objects/81/\n.d..t...... src/github.com/osism/testbed/.git/objects/82/\n.d..t...... src/github.com/osism/testbed/.git/objects/83/\n.d..t...... src/github.com/osism/testbed/.git/objects/84/\n.d..t...... src/github.com/osism/testbed/.git/objects/85/\n.d..t...... src/github.com/osism/testbed/.git/objects/86/\n.d..t...... src/github.com/osism/testbed/.git/objects/87/\n.d..t...... src/github.com/osism/testbed/.git/objects/8a/\n.d..t...... src/github.com/osism/testbed/.git/objects/8b/\n.d..t...... src/github.com/osism/testbed/.git/objects/8c/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/8c/58be1c5ed56d4f859eb72ab64b28c09731e1ef\n.d..t...... src/github.com/osism/testbed/.git/objects/8e/\n.d..t...... src/github.com/osism/testbed/.git/objects/8f/\n.d..t...... src/github.com/osism/testbed/.git/objects/90/\n.d..t...... src/github.com/osism/testbed/.git/objects/91/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/91/2c8319527490a42fe290d694c824bf4bf1d600\n.d..t...... src/github.com/osism/testbed/.git/objects/92/\n.d..t...... src/github.com/osism/testbed/.git/objects/94/\n.d..t...... src/github.com/osism/testbed/.git/objects/95/\n.d..t...... src/github.com/osism/testbed/.git/objects/96/\n.d..t...... src/github.com/osism/testbed/.git/objects/97/\n.d..t...... src/github.com/osism/testbed/.git/objects/98/\n.d..t...... src/github.com/osism/testbed/.git/objects/9a/\n.d..t...... src/github.com/osism/testbed/.git/objects/9c/\n.d..t...... src/github.com/osism/testbed/.git/objects/9e/\n.d..t...... src/github.com/osism/testbed/.git/objects/9f/\n.d..t...... src/github.com/osism/testbed/.git/objects/a0/\n.d..t...... src/github.com/osism/testbed/.git/objects/a1/\n.d..t...... src/github.com/osism/testbed/.git/objects/a2/\n.d..t...... src/github.com/osism/testbed/.git/objects/a4/\n.d..t...... src/github.com/osism/testbed/.git/objects/a5/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/a5/88fab15dc2873f358d9d03911fcbedf2b972fe\n<f+++++++++ src/github.com/osism/testbed/.git/objects/a5/d74130585e103779b999487d1872e59df81609\n.d..t...... src/github.com/osism/testbed/.git/objects/a6/\n.d..t...... src/github.com/osism/testbed/.git/objects/a7/\n.d..t...... src/github.com/osism/testbed/.git/objects/a8/\n.d..t...... src/github.com/osism/testbed/.git/objects/a9/\n.d..t...... src/github.com/osism/testbed/.git/objects/aa/\n.d..t...... src/github.com/osism/testbed/.git/objects/ab/\n.d..t...... src/github.com/osism/testbed/.git/objects/ac/\n.d..t...... src/github.com/osism/testbed/.git/objects/ad/\n.d..t...... src/github.com/osism/testbed/.git/objects/ae/\n.d..t...... src/github.com/osism/testbed/.git/objects/af/\n.d..t...... src/github.com/osism/testbed/.git/objects/b0/\n.d..t...... src/github.com/osism/testbed/.git/objects/b1/\n.d..t...... src/github.com/osism/testbed/.git/objects/b3/\n.d..t...... src/github.com/osism/testbed/.git/objects/b4/\n.d..t...... src/github.com/osism/testbed/.git/objects/b5/\n.d..t...... src/github.com/osism/testbed/.git/objects/b6/\n.d..t...... src/github.com/osism/testbed/.git/objects/b8/\n.d..t...... src/github.com/osism/testbed/.git/objects/b9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ba/\n.d..t...... src/github.com/osism/testbed/.git/objects/bc/\n.d..t...... src/github.com/osism/testbed/.git/objects/bd/\n.d..t...... src/github.com/osism/testbed/.git/objects/be/\n.d..t...... src/github.com/osism/testbed/.git/objects/bf/\n.d..t...... src/github.com/osism/testbed/.git/objects/c0/\n.d..t...... src/github.com/osism/testbed/.git/objects/c1/\n.d..t...... src/github.com/osism/testbed/.git/objects/c2/\n.d..t...... src/github.com/osism/testbed/.git/objects/c3/\n.d..t...... src/github.com/osism/testbed/.git/objects/c5/\n.d..t...... src/github.com/osism/testbed/.git/objects/c6/\n.d..t...... src/github.com/osism/testbed/.git/objects/c7/\n.d..t...... src/github.com/osism/testbed/.git/objects/c8/\n.d..t...... src/github.com/osism/testbed/.git/objects/c9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ca/\n.d..t...... src/github.com/osism/testbed/.git/objects/cb/\n.d..t...... src/github.com/osism/testbed/.git/objects/cc/\n.d..t...... src/github.com/osism/testbed/.git/objects/ce/\n.d..t...... src/github.com/osism/testbed/.git/objects/cf/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/cf/f6283404420c22109333470877dbb68d20f1e1\n.d..t...... src/github.com/osism/testbed/.git/objects/d0/\n.d..t...... src/github.com/osism/testbed/.git/objects/d2/\n.d..t...... src/github.com/osism/testbed/.git/objects/d3/\n.d..t...... src/github.com/osism/testbed/.git/objects/d5/\n.d..t...... src/github.com/osism/testbed/.git/objects/d6/\n.d..t...... src/github.com/osism/testbed/.git/objects/d7/\n.d..t...... src/github.com/osism/testbed/.git/objects/d9/\n.d..t...... src/github.com/osism/testbed/.git/objects/da/\n.d..t...... src/github.com/osism/testbed/.git/objects/db/\n.d..t...... src/github.com/osism/testbed/.git/objects/dc/\n.d..t...... src/github.com/osism/testbed/.git/objects/dd/\n.d..t...... src/github.com/osism/testbed/.git/objects/de/\n.d..t...... src/github.com/osism/testbed/.git/objects/df/\n.d..t...... src/github.com/osism/testbed/.git/objects/e0/\n.d..t...... src/github.com/osism/testbed/.git/objects/e2/\n.d..t...... src/github.com/osism/testbed/.git/objects/e3/\n.d..t...... src/github.com/osism/testbed/.git/objects/e4/\n.d..t...... src/github.com/osism/testbed/.git/objects/e5/\n.d..t...... src/github.com/osism/testbed/.git/objects/e6/\n.d..t...... src/github.com/osism/testbed/.git/objects/e7/\n.d..t...... src/github.com/osism/testbed/.git/objects/e8/\n.d..t...... src/github.com/osism/testbed/.git/objects/e9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ea/\n.d..t...... src/github.com/osism/testbed/.git/objects/eb/\n.d..t...... src/github.com/osism/testbed/.git/objects/ec/\n.d..t...... src/github.com/osism/testbed/.git/objects/ed/\n.d..t...... src/github.com/osism/testbed/.git/objects/ee/\n.d..t...... src/github.com/osism/testbed/.git/objects/ef/\n.d..t...... src/github.com/osism/testbed/.git/objects/f0/\n.d..t...... src/github.com/osism/testbed/.git/objects/f1/\n<f+++++++++ src/github.com/osism/testbed/.git/objects/f1/a4d73000ed828340ad8245bc13b541c3fb39c3\n.d..t...... src/github.com/osism/testbed/.git/objects/f2/\n.d..t...... src/github.com/osism/testbed/.git/objects/f3/\n.d..t...... src/github.com/osism/testbed/.git/objects/f5/\n.d..t...... src/github.com/osism/testbed/.git/objects/f8/\n.d..t...... src/github.com/osism/testbed/.git/objects/f9/\n.d..t...... src/github.com/osism/testbed/.git/objects/fa/\n.d..t...... src/github.com/osism/testbed/.git/objects/fb/\n.d..t...... src/github.com/osism/testbed/.git/objects/fc/\n.d..t...... src/github.com/osism/testbed/.git/objects/fd/\n.d..t...... src/github.com/osism/testbed/.git/objects/fe/\n.d..t...... src/github.com/osism/testbed/.git/objects/ff/\n.d..t...... src/github.com/osism/testbed/.git/objects/info/\n.d..t...... src/github.com/osism/testbed/.git/objects/pack/\n.d..t...... src/github.com/osism/testbed/.git/refs/\n.d..t...... src/github.com/osism/testbed/.github/\n<f..t...... src/github.com/osism/testbed/.github/renovate.json\n.d..t...... src/github.com/osism/testbed/.github/workflows/\n<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml\n.d..t...... src/github.com/osism/testbed/ansible/\n<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml\n<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml\n.d..t...... src/github.com/osism/testbed/containerlab/\n<f..t...... src/github.com/osism/testbed/containerlab/README.md\n<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml\n.d..t...... src/github.com/osism/testbed/containerlab/files/\n<f..t...... src/github.com/osism/testbed/containerlab/files/daemons\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf\n.d..t...... src/github.com/osism/testbed/contrib/\n<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py\n<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql\n<f..t...... src/github.com/osism/testbed/contrib/semver2.sh\n<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py\n.d..t...... src/github.com/osism/testbed/contrib/ansible/\n<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml\n<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample\n.d..t...... src/github.com/osism/testbed/contrib/assets/\n<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot\n.d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster\n.d..t...... src/github.com/osism/testbed/contrib/openstack-cli/\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example\n.d..t...... src/github.com/osism/testbed/contrib/ownca/\n<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml\n.d..t...... src/github.com/osism/testbed/environments/\n<f..t...... src/github.com/osism/testbed/environments/.vault_pass\n<f..t...... src/github.com/osism/testbed/environments/ansible.cfg\n<f..t...... src/github.com/osism/testbed/environments/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/images.yml\n<f..t...... src/github.com/osism/testbed/environments/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph.test/\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph/\n<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/\n<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/files/\n<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact\n.d..t...... src/github.com/osism/testbed/environments/custom/templates/\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2\n.d..t...... src/github.com/osism/testbed/environments/generic/\n<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/images.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/\n<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/files/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/\n<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/hosts\n<f..t...... src/github.com/osism/testbed/environments/manager/images.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/files/\n<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml\n.d..t...... src/github.com/osism/testbed/environments/monitoring/\n<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/openstack/\n<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub\n<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/openrc\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible-iaas.lst\n.d..t...... src/github.com/osism/testbed/inventory/\n<f..t...... src/github.com/osism/testbed/inventory/10-custom\n<f..t...... src/github.com/osism/testbed/inventory/20-roles\n<f..t...... src/github.com/osism/testbed/inventory/99-overwrite\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/all/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml\n.d..t...... src/github.com/osism/testbed/netbox/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml\n.d..t...... src/github.com/osism/testbed/netbox/moduletypes/\n<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep\n.d..t...... src/github.com/osism/testbed/netbox/resources/\n<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml\n.d..t...... src/github.com/osism/testbed/network/\n<f..t...... src/github.com/osism/testbed/network/iptables.sh\n<f..t...... src/github.com/osism/testbed/network/vxlan.sh\n.d..t...... src/github.com/osism/testbed/playbooks/\n<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml\n<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml\n<f..t...... src/github.com/osism/testbed/playbooks/post.yml\n<f..t...... src/github.com/osism/testbed/playbooks/pre.yml\n<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml\n.d..t...... src/github.com/osism/testbed/playbooks/templates/\n<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2\n.d..t...... src/github.com/osism/testbed/playbooks/vars/\n<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml\n<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml\n.d..t...... src/github.com/osism/testbed/scripts/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/check.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh\n<f..t...... src/github.com/osism/testbed/scripts/include.sh\n<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh\n<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh\n.d..t...... src/github.com/osism/testbed/scripts/bootstrap/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh\n.d..t...... src/github.com/osism/testbed/scripts/check/\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh\n<f+++++++++ src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible-iaas.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh\n.d..t...... src/github.com/osism/testbed/scripts/customisations/\n<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh\n.d..t...... src/github.com/osism/testbed/scripts/deploy/\n<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/scripts/upgrade/\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/terraform/\n<f..t...... src/github.com/osism/testbed/terraform/Makefile\n<f..t...... src/github.com/osism/testbed/terraform/README.md\n<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample\n<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/local.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample\n.d..t...... src/github.com/osism/testbed/terraform/environments/\n<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars\n.d..t...... src/github.com/osism/testbed/terraform/scripts/\n<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/\n*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/terraform-provider-local\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/environment\n*deleting   src/github.com/osism/testbed/terraform/.terraform/\n*deleting   src/github.com/osism/testbed/terraform/variables.tf\n*deleting   src/github.com/osism/testbed/terraform/provider.tf\n*deleting   src/github.com/osism/testbed/terraform/outputs.tf\n*deleting   src/github.com/osism/testbed/terraform/nova.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron.tf\n*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/manager.tf\n*deleting   src/github.com/osism/testbed/terraform/main.tf\n*deleting   src/github.com/osism/testbed/terraform/local.env\n*deleting   src/github.com/osism/testbed/terraform/inventory.ci\n*deleting   src/github.com/osism/testbed/terraform/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/data.tf\n*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars\n*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/.deploy.ci\n*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci\n",
                            "rc": 0,
                            "stdout_lines": [
                                ".d..t...... src/",
                                ".d..t...... src/github.com/",
                                ".d..t...... src/github.com/osism/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-services/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/47/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/65/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/92/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2",
                                ".d..t...... src/github.com/osism/terraform-base/",
                                "<f..t...... src/github.com/osism/terraform-base/.yamllint.yml",
                                "<f..t...... src/github.com/osism/terraform-base/.zuul.yaml",
                                "<f..t...... src/github.com/osism/terraform-base/LICENSE",
                                "<f..t...... src/github.com/osism/terraform-base/README.md",
                                ".d..t...... src/github.com/osism/terraform-base/.git/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config.worktree",
                                "<f..t...... src/github.com/osism/terraform-base/.git/description",
                                "<f..t...... src/github.com/osism/terraform-base/.git/index",
                                "<f..t...... src/github.com/osism/terraform-base/.git/packed-refs",
                                ".d..t...... src/github.com/osism/terraform-base/.git/branches/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/hooks/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/terraform-base/.git/info/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/info/exclude",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/00/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/01/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/02/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/06/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/07/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/08/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/10/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/12/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/15/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/16/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/17/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/20/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/21/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/24/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/26/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/36/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/37/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/41/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/43/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/46/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/47/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/48/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/49/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/52/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/55/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/56/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/58/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/61/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/63/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/66/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/67/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/68/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/70/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/72/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/81/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/82/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/83/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/84/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/85/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/86/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/88/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/90/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/91/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/92/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/95/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/96/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/98/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/af/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/da/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/db/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/de/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/info/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.github/",
                                "<f..t...... src/github.com/osism/terraform-base/.github/renovate.json",
                                ".d..t...... src/github.com/osism/terraform-base/playbooks/",
                                "<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/environments/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf",
                                ".d..t...... src/github.com/osism/testbed/",
                                "<f..t...... src/github.com/osism/testbed/.ansible-lint",
                                "<f..t...... src/github.com/osism/testbed/.flake8",
                                "<f..t...... src/github.com/osism/testbed/.gitignore",
                                "<f..t...... src/github.com/osism/testbed/.yamllint.yml",
                                "<f..t...... src/github.com/osism/testbed/.zuul.yaml",
                                "<f..t...... src/github.com/osism/testbed/LICENSE",
                                "<f..t...... src/github.com/osism/testbed/Makefile",
                                "<f..t...... src/github.com/osism/testbed/README.md",
                                "<f..t...... src/github.com/osism/testbed/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/gilt.yml",
                                "<f..t...... src/github.com/osism/testbed/requirements.txt",
                                "<f..t...... src/github.com/osism/testbed/requirements.yml",
                                "<f..t...... src/github.com/osism/testbed/tox.ini",
                                ".d..t...... src/github.com/osism/testbed/.git/",
                                "<f..t...... src/github.com/osism/testbed/.git/HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/config",
                                "<f..t...... src/github.com/osism/testbed/.git/config.worktree",
                                "<f..t...... src/github.com/osism/testbed/.git/description",
                                "<f.st...... src/github.com/osism/testbed/.git/index",
                                "<f..t...... src/github.com/osism/testbed/.git/packed-refs",
                                ".d..t...... src/github.com/osism/testbed/.git/branches/",
                                ".d..t...... src/github.com/osism/testbed/.git/hooks/",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/testbed/.git/info/",
                                "<f..t...... src/github.com/osism/testbed/.git/info/exclude",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/00/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/01/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/02/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/03/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/04/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/05/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/05/c88d3bc61d80be6a0190ba4158bc7856df0538",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/06/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/07/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/08/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/09/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/10/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/11/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/12/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/14/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/15/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/16/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/17/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/18/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/19/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/20/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/20/0c611c75f445983f13b8da3e93db7d47ed81bb",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/21/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/22/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/23/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/24/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/25/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/26/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/27/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/28/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/30/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/31/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/31/bd0867a2edfe0f397bd15bb36566a2cd7fd027",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/32/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/33/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/34/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/35/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/35/7756bf2a17acc26a5f1f6c2d0a36447a7e80eb",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/36/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/37/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/38/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/39/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3d/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/3d/34dfba029d7b86d4254f892fb6ac8d456e32de",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/40/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/40/46222976dc495717131ef332b39bc7929e3001",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/41/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/42/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/42/b694dcc3f1dae4f49ac7840bff7c0ccc770b41",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/43/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/44/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/45/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/46/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/47/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/48/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/49/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4d/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/4d/515798706b72a52536f96fd021877023a4df22",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4e/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/4e/f4687bc0e265f5692b55f6fb3b7e1ecb661918",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/50/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/51/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/52/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/54/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/55/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/57/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/58/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5f/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/5f/08648a61449abc0a5fd129393ef87b0b64902b",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/60/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/61/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/64/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/66/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/68/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/69/",
                                "cd+++++++++ src/github.com/osism/testbed/.git/objects/6a/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/6a/98d0e99dc7e391ce46704852e375bdc6ee2406",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/70/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/70/4ea3b380be6639d26c68e230e4acadd1b36f78",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/71/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/72/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/73/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/75/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/76/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/77/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/78/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/78/f8e688db2b72f61840ef15817a13b831e2e64d",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/79/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/80/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/81/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/82/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/83/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/84/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/85/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/86/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/87/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8c/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/8c/58be1c5ed56d4f859eb72ab64b28c09731e1ef",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/90/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/91/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/91/2c8319527490a42fe290d694c824bf4bf1d600",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/92/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/94/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/95/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/96/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/97/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/98/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a5/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/a5/88fab15dc2873f358d9d03911fcbedf2b972fe",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/a5/d74130585e103779b999487d1872e59df81609",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/af/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/be/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cf/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/cf/f6283404420c22109333470877dbb68d20f1e1",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/da/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/db/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/de/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/df/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f1/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/objects/f1/a4d73000ed828340ad8245bc13b541c3fb39c3",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/info/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/testbed/.git/refs/",
                                ".d..t...... src/github.com/osism/testbed/.github/",
                                "<f..t...... src/github.com/osism/testbed/.github/renovate.json",
                                ".d..t...... src/github.com/osism/testbed/.github/workflows/",
                                "<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml",
                                ".d..t...... src/github.com/osism/testbed/ansible/",
                                "<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/README.md",
                                "<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/files/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/daemons",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf",
                                ".d..t...... src/github.com/osism/testbed/contrib/",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py",
                                "<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql",
                                "<f..t...... src/github.com/osism/testbed/contrib/semver2.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py",
                                ".d..t...... src/github.com/osism/testbed/contrib/ansible/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample",
                                ".d..t...... src/github.com/osism/testbed/contrib/assets/",
                                "<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot",
                                ".d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster",
                                ".d..t...... src/github.com/osism/testbed/contrib/openstack-cli/",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example",
                                ".d..t...... src/github.com/osism/testbed/contrib/ownca/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/",
                                "<f..t...... src/github.com/osism/testbed/environments/.vault_pass",
                                "<f..t...... src/github.com/osism/testbed/environments/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/environments/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph.test/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/templates/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2",
                                ".d..t...... src/github.com/osism/testbed/environments/generic/",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/hosts",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/monitoring/",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/openrc",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible-iaas.lst",
                                ".d..t...... src/github.com/osism/testbed/inventory/",
                                "<f..t...... src/github.com/osism/testbed/inventory/10-custom",
                                "<f..t...... src/github.com/osism/testbed/inventory/20-roles",
                                "<f..t...... src/github.com/osism/testbed/inventory/99-overwrite",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/all/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/moduletypes/",
                                "<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/netbox/resources/",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml",
                                ".d..t...... src/github.com/osism/testbed/network/",
                                "<f..t...... src/github.com/osism/testbed/network/iptables.sh",
                                "<f..t...... src/github.com/osism/testbed/network/vxlan.sh",
                                ".d..t...... src/github.com/osism/testbed/playbooks/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/post.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/pre.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml",
                                ".d..t...... src/github.com/osism/testbed/playbooks/templates/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2",
                                ".d..t...... src/github.com/osism/testbed/playbooks/vars/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml",
                                ".d..t...... src/github.com/osism/testbed/scripts/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/include.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/bootstrap/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/check/",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh",
                                "<f+++++++++ src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible-iaas.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/customisations/",
                                "<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/deploy/",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/upgrade/",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/terraform/",
                                "<f..t...... src/github.com/osism/testbed/terraform/Makefile",
                                "<f..t...... src/github.com/osism/testbed/terraform/README.md",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/local.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample",
                                ".d..t...... src/github.com/osism/testbed/terraform/environments/",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars",
                                ".d..t...... src/github.com/osism/testbed/terraform/scripts/",
                                "<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/terraform-provider-local",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.6.1/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/environment",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/",
                                "*deleting   src/github.com/osism/testbed/terraform/variables.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/provider.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/outputs.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nova.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/main.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/local.env",
                                "*deleting   src/github.com/osism/testbed/terraform/inventory.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/data.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars",
                                "*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.deploy.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-b1ff-a440-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:09.132306Z",
                            "start": "2026-01-13T00:00:48.422636Z"
                        },
                        "id": "0242ac1f-6606-b1ff-a440-00000000002d",
                        "name": "Synchronize src repos to workspace directory."
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 9,
            "failures": 0,
            "ignored": 0,
            "ok": 26,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/osism/testbed/playbooks/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T00:02:18.090530Z",
                    "start": "2026-01-13T00:02:10.641938Z"
                },
                "id": "0242ac1f-6606-05a6-3277-000000000002",
                "name": "Pre play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:10.723641Z",
                            "start": "2026-01-13T00:02:10.661147Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:10.781276Z",
                            "start": "2026-01-13T00:02:10.730736Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.484101",
                            "end": "2026-01-13 00:02:13.242974",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-05a6-3277-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:02:11.758873",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-01-13 00:02:12 - clean up servers\n2026-01-13 00:02:12 - clean up keypairs\n2026-01-13 00:02:12 - wait for servers to be gone\n2026-01-13 00:02:12 - clean up ports\n2026-01-13 00:02:12 - clean up volumes\n2026-01-13 00:02:12 - disconnect routers\n2026-01-13 00:02:12 - clean up subnets\n2026-01-13 00:02:12 - clean up networks\n2026-01-13 00:02:13 - clean up security groups\n2026-01-13 00:02:13 - clean up floating ips\n2026-01-13 00:02:13 - clean up routers",
                            "stdout_lines": [
                                "2026-01-13 00:02:12 - clean up servers",
                                "2026-01-13 00:02:12 - clean up keypairs",
                                "2026-01-13 00:02:12 - wait for servers to be gone",
                                "2026-01-13 00:02:12 - clean up ports",
                                "2026-01-13 00:02:12 - clean up volumes",
                                "2026-01-13 00:02:12 - disconnect routers",
                                "2026-01-13 00:02:12 - clean up subnets",
                                "2026-01-13 00:02:12 - clean up networks",
                                "2026-01-13 00:02:13 - clean up security groups",
                                "2026-01-13 00:02:13 - clean up floating ips",
                                "2026-01-13 00:02:13 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-05a6-3277-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:13.322245Z",
                            "start": "2026-01-13T00:02:10.822950Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000006",
                        "name": "Clean the cloud environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.unarchive",
                            "changed": false,
                            "dest": "/home/zuul-testbed05",
                            "gid": 1006,
                            "group": "zuul-testbed05",
                            "handler": "ZipArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": null,
                                    "decrypt": true,
                                    "dest": "/home/zuul-testbed05",
                                    "exclude": [],
                                    "extra_opts": [],
                                    "group": null,
                                    "include": [
                                        "tofu"
                                    ],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/opentofu/opentofu/releases/download/v1.11.2/tofu_1.11.2_linux_amd64.zip",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed05",
                            "size": 4096,
                            "src": "/home/zuul-testbed05/.ansible/tmp/ansible-tmp-1768262533.3966815-13-29362029955175/tofu_1.11.2_linux_amd64w7eltiwx.zip",
                            "state": "directory",
                            "uid": 1005
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:16.504961Z",
                            "start": "2026-01-13T00:02:13.334369Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000007",
                        "name": "Extract tofu binary"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "rsync",
                                "-avz",
                                "/home/zuul-testbed05/src/github.com/osism/terraform-base/testbed-default/",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform"
                            ],
                            "delta": "0:00:00.057102",
                            "end": "2026-01-13 00:02:16.891385",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "rsync -avz /home/zuul-testbed05/src/github.com/osism/terraform-base/testbed-default/ /home/zuul-testbed05/src/github.com/osism/testbed/terraform",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-05a6-3277-000000000008-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:02:16.834283",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sending incremental file list\ndata.tf\nmain.tf\nmanager.tf\nneutron.tf\nnodes.tf\nnova.tf\noutputs.tf\nprovider.tf\nvariables.tf\ncustomisations/\ncustomisations/access_floatingip_custom.tf\ncustomisations/access_ipv4_custom.tf\ncustomisations/access_ipv6_custom.tf\ncustomisations/default_custom.tf\ncustomisations/external_api_custom.tf\ncustomisations/neutron_floatingip_custom.tf\noverrides/\noverrides/manager_boot_from_image_override.tf\noverrides/manager_boot_from_volume_override.tf\noverrides/neutron_availability_zone_hints_network_override.tf\noverrides/neutron_availability_zone_hints_router_override.tf\noverrides/neutron_router_enable_snat_override.tf\noverrides/nodes_boot_from_image_override.tf\noverrides/nodes_boot_from_volume_override.tf\noverrides/nodes_use_ephemeral_storage_override.tf\n\nsent 6,900 bytes  received 470 bytes  14,740.00 bytes/sec\ntotal size is 26,532  speedup is 3.60",
                            "stdout_lines": [
                                "sending incremental file list",
                                "data.tf",
                                "main.tf",
                                "manager.tf",
                                "neutron.tf",
                                "nodes.tf",
                                "nova.tf",
                                "outputs.tf",
                                "provider.tf",
                                "variables.tf",
                                "customisations/",
                                "customisations/access_floatingip_custom.tf",
                                "customisations/access_ipv4_custom.tf",
                                "customisations/access_ipv6_custom.tf",
                                "customisations/default_custom.tf",
                                "customisations/external_api_custom.tf",
                                "customisations/neutron_floatingip_custom.tf",
                                "overrides/",
                                "overrides/manager_boot_from_image_override.tf",
                                "overrides/manager_boot_from_volume_override.tf",
                                "overrides/neutron_availability_zone_hints_network_override.tf",
                                "overrides/neutron_availability_zone_hints_router_override.tf",
                                "overrides/neutron_router_enable_snat_override.tf",
                                "overrides/nodes_boot_from_image_override.tf",
                                "overrides/nodes_boot_from_volume_override.tf",
                                "overrides/nodes_use_ephemeral_storage_override.tf",
                                "",
                                "sent 6,900 bytes  received 470 bytes  14,740.00 bytes/sec",
                                "total size is 26,532  speedup is 3.60"
                            ],
                            "zuul_log_id": "0242ac1f-6606-05a6-3277-000000000008-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:17.072772Z",
                            "start": "2026-01-13T00:02:16.555981Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000008",
                        "name": "Sync terraform blueprint"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.template",
                            "changed": true,
                            "checksum": "e03860e44270132bd1bf7d494e1d7de81ec3d2ac",
                            "dest": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/local.env",
                            "diff": [],
                            "gid": 1006,
                            "group": "zuul-testbed05",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "local.env.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "e03860e44270132bd1bf7d494e1d7de81ec3d2ac",
                                    "content": null,
                                    "dest": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/local.env",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": "0644",
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul-testbed05/.ansible/tmp/ansible-tmp-1768262537.1716962-22-227838678746142/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "49f96bcad982355361b2398e1d0fd032",
                            "mode": "0644",
                            "owner": "zuul-testbed05",
                            "size": 42,
                            "src": "/home/zuul-testbed05/.ansible/tmp/ansible-tmp-1768262537.1716962-22-227838678746142/source",
                            "state": "file",
                            "uid": 1005
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:18.090530Z",
                            "start": "2026-01-13T00:02:17.084673Z"
                        },
                        "id": "0242ac1f-6606-05a6-3277-000000000009",
                        "name": "Create local.env file"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 3,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/osism/testbed/playbooks/deploy.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:40:38.375052Z",
                    "start": "2026-01-13T00:02:19.162198Z"
                },
                "id": "0242ac1f-6606-8e06-7a3c-000000000002",
                "name": "Deploy testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:19.282333Z",
                            "start": "2026-01-13T00:02:19.192929Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000004",
                        "name": "Print used ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version": "2024.2",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:19.353425Z",
                            "start": "2026-01-13T00:02:19.289667Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000005",
                        "name": "Print used openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version": "latest",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:19.425032Z",
                            "start": "2026-01-13T00:02:19.363245Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000006",
                        "name": "Print used manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "image_username": "ubuntu"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:19.518234Z",
                            "start": "2026-01-13T00:02:19.434293Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000007",
                        "name": "Set facts (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:02:19.562258Z",
                            "start": "2026-01-13T00:02:19.525198Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000008",
                        "name": "Set facts (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "make",
                                "ENVIRONMENT=ci-ubuntu-24.04",
                                "CLOUD=ci",
                                "IMAGE_USERNAME=ubuntu",
                                "TERRAFORM=/home/zuul-testbed05/tofu",
                                "VERSION_CEPH=reef",
                                "VERSION_MANAGER=latest",
                                "VERSION_OPENSTACK=2024.2",
                                "TEMPEST=true",
                                "CEPH_STACK=ceph-ansible",
                                "create"
                            ],
                            "delta": "0:01:20.334471",
                            "end": "2026-01-13 00:03:40.768860",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "make\n  ENVIRONMENT=ci-ubuntu-24.04\n  CLOUD=ci\n  IMAGE_USERNAME=ubuntu\n  TERRAFORM=/home/zuul-testbed05/tofu\n  VERSION_CEPH=reef\n  VERSION_MANAGER=latest\n  VERSION_OPENSTACK=2024.2\n  TEMPEST=true\n  CEPH_STACK=ceph-ansible\n  create\n",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:02:20.434389",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\nInitializing the backend...\n\nInitializing provider plugins...\n- terraform.io/builtin/terraform is built in to OpenTofu\n- Finding latest version of hashicorp/null...\n- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...\n- Finding hashicorp/local versions matching \">= 2.2.0\"...\n- Installing hashicorp/null v3.2.4...\n- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)\n- Installing terraform-provider-openstack/openstack v3.4.0...\n- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)\n- Installing hashicorp/local v2.6.1...\n- Installed hashicorp/local v2.6.1 (signed, key ID 0C0AF313E5FD9F80)\n\nProviders are signed by their developers.\nIf you'd like to know more about provider signing, you can read about it here:\nhttps://opentofu.org/docs/cli/plugins/signing/\n\nOpenTofu has created a lock file .terraform.lock.hcl to record the provider\nselections it made above. Include this file in your version control repository\nso that OpenTofu can guarantee to make the same selections by default when\nyou run \"tofu init\" in the future.\n\nOpenTofu has been successfully initialized!\n\nYou may now begin working with OpenTofu. Try running \"tofu plan\" to see\nany changes that are required for your infrastructure. All OpenTofu commands\nshould now work.\n\nIf you ever set or change modules or backend configuration for OpenTofu,\nrerun this command to reinitialize your working directory. If you forget, other\ncommands will detect it and remind you to do so if necessary.\nCreated and switched to workspace \"ci\"!\n\nYou're now on a new, empty workspace. Workspaces isolate their state,\nso if you run \"tofu plan\" OpenTofu will not see any existing state\nfor this configuration.\nci.auto.tfvars\ndefault_custom.tf\ndata.openstack_networking_network_v2.public: Reading...\ndata.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]\n\nOpenTofu used the selected providers to generate the following execution\nplan. Resource actions are indicated with the following symbols:\n  + create\n <= read (data resources)\n\nOpenTofu will perform the following actions:\n\n  # data.openstack_images_image_v2.image will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # data.openstack_images_image_v2.image_node will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image_node\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # local_file.MANAGER_ADDRESS will be created\n  + resource \"local_file\" \"MANAGER_ADDRESS\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".MANAGER_ADDRESS.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.id_rsa_pub will be created\n  + resource \"local_file\" \"id_rsa_pub\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".id_rsa.ci.pub\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.inventory will be created\n  + resource \"local_file\" \"inventory\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \"inventory.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_sensitive_file.id_rsa will be created\n  + resource \"local_sensitive_file\" \"id_rsa\" {\n      + content              = (sensitive value)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0700\"\n      + file_permission      = \"0600\"\n      + filename             = \".id_rsa.ci\"\n      + id                   = (known after apply)\n    }\n\n  # null_resource.node_semaphore will be created\n  + resource \"null_resource\" \"node_semaphore\" {\n      + id = (known after apply)\n    }\n\n  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-manager-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[6] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-6-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[7] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-7-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[8] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-8-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_compute_instance_v2.manager_server will be created\n  + resource \"openstack_compute_instance_v2\" \"manager_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-4V-16\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-manager\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = (sensitive value)\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[0] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-0\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[1] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-1\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[2] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-2\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[3] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-3\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[4] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-4\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[5] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-5\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_keypair_v2.key will be created\n  + resource \"openstack_compute_keypair_v2\" \"key\" {\n      + fingerprint = (known after apply)\n      + id          = (known after apply)\n      + name        = \"testbed\"\n      + private_key = (sensitive value)\n      + public_key  = (known after apply)\n      + region      = (known after apply)\n      + user_id     = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created\n  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {\n      + fixed_ip    = (known after apply)\n      + floating_ip = (known after apply)\n      + id          = (known after apply)\n      + port_id     = (known after apply)\n      + region      = (known after apply)\n    }\n\n  # openstack_networking_floatingip_v2.manager_floating_ip will be created\n  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {\n      + address    = (known after apply)\n      + all_tags   = (known after apply)\n      + dns_domain = (known after apply)\n      + dns_name   = (known after apply)\n      + fixed_ip   = (known after apply)\n      + id         = (known after apply)\n      + pool       = \"public\"\n      + port_id    = (known after apply)\n      + region     = (known after apply)\n      + subnet_id  = (known after apply)\n      + tenant_id  = (known after apply)\n    }\n\n  # openstack_networking_network_v2.net_management will be created\n  + resource \"openstack_networking_network_v2\" \"net_management\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + dns_domain              = (known after apply)\n      + external                = (known after apply)\n      + id                      = (known after apply)\n      + mtu                     = (known after apply)\n      + name                    = \"net-testbed-management\"\n      + port_security_enabled   = (known after apply)\n      + qos_policy_id           = (known after apply)\n      + region                  = (known after apply)\n      + shared                  = (known after apply)\n      + tenant_id               = (known after apply)\n      + transparent_vlan        = (known after apply)\n\n      + segments (known after apply)\n    }\n\n  # openstack_networking_port_v2.manager_port_management will be created\n  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.5\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[0] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.10\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[1] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.11\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[2] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.12\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[3] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.13\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[4] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.14\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[5] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.15\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_router_interface_v2.router_interface will be created\n  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {\n      + force_destroy = false\n      + id            = (known after apply)\n      + port_id       = (known after apply)\n      + region        = (known after apply)\n      + router_id     = (known after apply)\n      + subnet_id     = (known after apply)\n    }\n\n  # openstack_networking_router_v2.router will be created\n  + resource \"openstack_networking_router_v2\" \"router\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + distributed             = (known after apply)\n      + enable_snat             = (known after apply)\n      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"\n      + external_qos_policy_id  = (known after apply)\n      + id                      = (known after apply)\n      + name                    = \"testbed\"\n      + region                  = (known after apply)\n      + tenant_id               = (known after apply)\n\n      + external_fixed_ip (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {\n      + description             = \"ssh\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 22\n      + port_range_min          = 22\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {\n      + description             = \"wireguard\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 51820\n      + port_range_min          = 51820\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {\n      + description             = \"vrrp\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"112\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_management will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {\n      + all_tags    = (known after apply)\n      + description = \"management security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-management\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_node will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {\n      + all_tags    = (known after apply)\n      + description = \"node security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-node\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_subnet_v2.subnet_management will be created\n  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {\n      + all_tags          = (known after apply)\n      + cidr              = \"192.168.16.0/20\"\n      + dns_nameservers   = [\n          + \"8.8.8.8\",\n          + \"9.9.9.9\",\n        ]\n      + enable_dhcp       = true\n      + gateway_ip        = (known after apply)\n      + id                = (known after apply)\n      + ip_version        = 4\n      + ipv6_address_mode = (known after apply)\n      + ipv6_ra_mode      = (known after apply)\n      + name              = \"subnet-testbed-management\"\n      + network_id        = (known after apply)\n      + no_gateway        = false\n      + region            = (known after apply)\n      + service_types     = (known after apply)\n      + tenant_id         = (known after apply)\n\n      + allocation_pool {\n          + end   = \"192.168.31.250\"\n          + start = \"192.168.31.200\"\n        }\n    }\n\n  # terraform_data.image will be created\n  + resource \"terraform_data\" \"image\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\n  # terraform_data.image_node will be created\n  + resource \"terraform_data\" \"image_node\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\nPlan: 64 to add, 0 to change, 0 to destroy.\n\nChanges to Outputs:\n  + manager_address = (sensitive value)\n  + private_key     = (sensitive value)\nterraform_data.image_node: Creating...\nterraform_data.image_node: Creation complete after 0s [id=e1c2e4bb-a477-452d-9478-bcd164c47803]\ndata.openstack_images_image_v2.image_node: Reading...\nopenstack_networking_network_v2.net_management: Creating...\nopenstack_compute_keypair_v2.key: Creating...\nopenstack_blockstorage_volume_v3.node_volume[7]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creating...\nterraform_data.image: Creating...\nterraform_data.image: Creation complete after 0s [id=d816786d-72d9-acb3-b0bb-a6b8f5fccd60]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creating...\ndata.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]\nopenstack_blockstorage_volume_v3.node_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creating...\nopenstack_networking_network_v2.net_management: Creation complete after 1s [id=ab1787c0-a6c8-4860-adff-4d1b12555b92]\ndata.openstack_images_image_v2.image: Reading...\ndata.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=f69e02e7-d854-4ded-bb8d-51d0e0400336]\nlocal_file.id_rsa_pub: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=79922d84-0445-4535-976b-32e74e35a748]\nopenstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=5295d09e-fddd-4452-8a25-9ba23e2b95ae]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creating...\nlocal_file.id_rsa_pub: Creation complete after 0s [id=7d94d81cdceac9b3b6345af92bb1be3add46cc11]\nopenstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=0a292857-8cd9-4a14-95ba-a5d022f4a90e]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 4s [id=49cd33e4-72cd-4f3f-940d-55c9f0f00a98]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=9db8234e-f6a8-4211-a809-87a509109e78]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 4s [id=5c0bff01-3898-4d25-903e-2ecdf087243c]\nlocal_sensitive_file.id_rsa: Creating...\nlocal_sensitive_file.id_rsa: Creation complete after 0s [id=aa1825b291d916f8a2c90b4b6acc8cc92ef06ae8]\nopenstack_networking_subnet_v2.subnet_management: Creating...\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 3s [id=1f00cc32-4927-4d99-9c1e-b649b1d1f573]\nopenstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=6ad71b9e-76db-4ac5-b372-050f59253056]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 4s [id=4dcaa69c-5414-4861-9f75-cc0da42200e7]\nopenstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=aeecf627-ee34-49a3-a604-f1ec5825b8f9]\nopenstack_networking_router_v2.router: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=83689365-4423-433a-82c0-63cbcaedfdf8]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=306cfbe9-242f-441d-bc49-37fa1b1f4569]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=5f6d3b65-3844-4001-8889-d6deb3f0644d]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=ffeaaf24-9754-44c8-bb36-eb3a5d2d5315]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=d36bd727-f6fd-4e09-af6c-5d1752a9fb11]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 3s [id=a3f90d2f-3417-480d-afb3-bab2acf5e837]\nopenstack_networking_router_v2.router: Creation complete after 4s [id=626bfd75-5d77-4996-991e-e4b06a8e0797]\nopenstack_networking_router_interface_v2.router_interface: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creation complete after 0s [id=82a0ffe4-c609-4bad-a240-ed41f4279f28]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...\nopenstack_networking_port_v2.manager_port_management: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=58bf471c-d80e-418b-b4d0-705ee3d0ea1c]\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 0s [id=449dd116-1fca-4840-b186-c61271871b7f]\nopenstack_networking_port_v2.node_port_management[5]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 0s [id=cd34c19d-4240-4fd7-abcf-9ef72006f683]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=828cb3f4-1eb7-49b3-b1b5-d9c7daab8229]\nopenstack_networking_port_v2.node_port_management[4]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=8353e7e2-b281-4ec9-9907-6ebbfc0f389b]\nopenstack_networking_port_v2.node_port_management[0]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 1s [id=19dacea6-f8ac-4dce-a10d-e93d240f1da4]\nopenstack_networking_port_v2.node_port_management[1]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 1s [id=0490a749-77b9-4f38-91ff-4b1321995ff2]\nopenstack_networking_port_v2.node_port_management[2]: Creating...\nopenstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=21f6ad4e-2daf-49cf-9f3a-83bd124ef2b6]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=3ca4a049-0969-4712-9f12-0f3f489ff07d]\nopenstack_networking_port_v2.node_port_management[3]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=f680df32-572e-4bf4-b0da-1405d80f137d]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 1s [id=43d3dec3-2b1f-4793-989e-a338aaadfc07]\nopenstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=5133aec3-bcc8-4613-a099-273522a85686]\nopenstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=033250be-3730-4e28-8bae-574aa81261e3]\nopenstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=b46751e7-e6e1-4843-8018-ec0fe26ce24c]\nopenstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=f9c66452-e535-44f2-abde-2653138d6638]\nopenstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=dc661a00-4fe6-413a-9683-d66479df831d]\nopenstack_networking_router_interface_v2.router_interface: Creation complete after 3s [id=e4c1491f-c59c-42b7-8ed3-e8c0ebd3203b]\nopenstack_networking_floatingip_v2.manager_floating_ip: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creation complete after 2s [id=93e66259-6123-483f-83e0-5ca60df99fb6]\nopenstack_compute_instance_v2.node_server[5]: Creating...\nopenstack_compute_instance_v2.node_server[3]: Creating...\nopenstack_compute_instance_v2.node_server[2]: Creating...\nopenstack_compute_instance_v2.node_server[4]: Creating...\nopenstack_compute_instance_v2.node_server[0]: Creating...\nopenstack_compute_instance_v2.node_server[1]: Creating...\nopenstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=1665fef5-3425-4ab0-90ee-a0f1212f09dc]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...\nlocal_file.MANAGER_ADDRESS: Creating...\nlocal_file.MANAGER_ADDRESS: Creation complete after 0s [id=01887e5a3def8b233d771971aedf02fc115bae67]\nlocal_file.inventory: Creating...\nlocal_file.inventory: Creation complete after 0s [id=b287740adb6f33ee390d5b061cc2099583ba72f7]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=1665fef5-3425-4ab0-90ee-a0f1212f09dc]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Creation complete after 31s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c]\nopenstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=71ad69d2-ecf0-4a82-9d7a-7b6f0722f356]\nopenstack_compute_instance_v2.node_server[1]: Creation complete after 31s [id=a714327d-623a-4d36-8acf-a2e00af61082]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Creation complete after 41s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b]\nopenstack_compute_instance_v2.node_server[4]: Creation complete after 41s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217]\nopenstack_compute_instance_v2.node_server[0]: Creation complete after 42s [id=57e5bb4e-daf6-427c-b4df-17c8f7f3e79e]\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...\nnull_resource.node_semaphore: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...\nnull_resource.node_semaphore: Creation complete after 0s [id=7303008649623477253]\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...\nopenstack_compute_instance_v2.manager_server: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 4s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/5295d09e-fddd-4452-8a25-9ba23e2b95ae]\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 4s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/0a292857-8cd9-4a14-95ba-a5d022f4a90e]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 4s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/f69e02e7-d854-4ded-bb8d-51d0e0400336]\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/9db8234e-f6a8-4211-a809-87a509109e78]\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 4s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/5c0bff01-3898-4d25-903e-2ecdf087243c]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 4s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/1f00cc32-4927-4d99-9c1e-b649b1d1f573]\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 10s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/49cd33e4-72cd-4f3f-940d-55c9f0f00a98]\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/79922d84-0445-4535-976b-32e74e35a748]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 10s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/6ad71b9e-76db-4ac5-b372-050f59253056]\nopenstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.manager_server: Creation complete after 21s [id=d3b2fcd4-dac4-415f-a170-928342d9466c]\n\nApply complete! Resources: 64 added, 0 changed, 0 destroyed.\n\nOutputs:\n\nmanager_address = <sensitive>\nprivate_key = <sensitive>",
                            "stdout_lines": [
                                "",
                                "Initializing the backend...",
                                "",
                                "Initializing provider plugins...",
                                "- terraform.io/builtin/terraform is built in to OpenTofu",
                                "- Finding latest version of hashicorp/null...",
                                "- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...",
                                "- Finding hashicorp/local versions matching \">= 2.2.0\"...",
                                "- Installing hashicorp/null v3.2.4...",
                                "- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing terraform-provider-openstack/openstack v3.4.0...",
                                "- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)",
                                "- Installing hashicorp/local v2.6.1...",
                                "- Installed hashicorp/local v2.6.1 (signed, key ID 0C0AF313E5FD9F80)",
                                "",
                                "Providers are signed by their developers.",
                                "If you'd like to know more about provider signing, you can read about it here:",
                                "https://opentofu.org/docs/cli/plugins/signing/",
                                "",
                                "OpenTofu has created a lock file .terraform.lock.hcl to record the provider",
                                "selections it made above. Include this file in your version control repository",
                                "so that OpenTofu can guarantee to make the same selections by default when",
                                "you run \"tofu init\" in the future.",
                                "",
                                "OpenTofu has been successfully initialized!",
                                "",
                                "You may now begin working with OpenTofu. Try running \"tofu plan\" to see",
                                "any changes that are required for your infrastructure. All OpenTofu commands",
                                "should now work.",
                                "",
                                "If you ever set or change modules or backend configuration for OpenTofu,",
                                "rerun this command to reinitialize your working directory. If you forget, other",
                                "commands will detect it and remind you to do so if necessary.",
                                "Created and switched to workspace \"ci\"!",
                                "",
                                "You're now on a new, empty workspace. Workspaces isolate their state,",
                                "so if you run \"tofu plan\" OpenTofu will not see any existing state",
                                "for this configuration.",
                                "ci.auto.tfvars",
                                "default_custom.tf",
                                "data.openstack_networking_network_v2.public: Reading...",
                                "data.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]",
                                "",
                                "OpenTofu used the selected providers to generate the following execution",
                                "plan. Resource actions are indicated with the following symbols:",
                                "  + create",
                                " <= read (data resources)",
                                "",
                                "OpenTofu will perform the following actions:",
                                "",
                                "  # data.openstack_images_image_v2.image will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # data.openstack_images_image_v2.image_node will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image_node\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.MANAGER_ADDRESS will be created",
                                "  + resource \"local_file\" \"MANAGER_ADDRESS\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".MANAGER_ADDRESS.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.id_rsa_pub will be created",
                                "  + resource \"local_file\" \"id_rsa_pub\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".id_rsa.ci.pub\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.inventory will be created",
                                "  + resource \"local_file\" \"inventory\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \"inventory.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_sensitive_file.id_rsa will be created",
                                "  + resource \"local_sensitive_file\" \"id_rsa\" {",
                                "      + content              = (sensitive value)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0700\"",
                                "      + file_permission      = \"0600\"",
                                "      + filename             = \".id_rsa.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # null_resource.node_semaphore will be created",
                                "  + resource \"null_resource\" \"node_semaphore\" {",
                                "      + id = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-manager-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[6] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-6-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[7] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-7-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[8] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-8-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.manager_server will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"manager_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-4V-16\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-manager\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = (sensitive value)",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[0] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-0\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[1] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-1\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[2] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-2\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[3] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-3\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[4] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-4\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[5] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-5\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_keypair_v2.key will be created",
                                "  + resource \"openstack_compute_keypair_v2\" \"key\" {",
                                "      + fingerprint = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed\"",
                                "      + private_key = (sensitive value)",
                                "      + public_key  = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + user_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created",
                                "  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {",
                                "      + fixed_ip    = (known after apply)",
                                "      + floating_ip = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + port_id     = (known after apply)",
                                "      + region      = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_v2.manager_floating_ip will be created",
                                "  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {",
                                "      + address    = (known after apply)",
                                "      + all_tags   = (known after apply)",
                                "      + dns_domain = (known after apply)",
                                "      + dns_name   = (known after apply)",
                                "      + fixed_ip   = (known after apply)",
                                "      + id         = (known after apply)",
                                "      + pool       = \"public\"",
                                "      + port_id    = (known after apply)",
                                "      + region     = (known after apply)",
                                "      + subnet_id  = (known after apply)",
                                "      + tenant_id  = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_network_v2.net_management will be created",
                                "  + resource \"openstack_networking_network_v2\" \"net_management\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + dns_domain              = (known after apply)",
                                "      + external                = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + mtu                     = (known after apply)",
                                "      + name                    = \"net-testbed-management\"",
                                "      + port_security_enabled   = (known after apply)",
                                "      + qos_policy_id           = (known after apply)",
                                "      + region                  = (known after apply)",
                                "      + shared                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "      + transparent_vlan        = (known after apply)",
                                "",
                                "      + segments (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.manager_port_management will be created",
                                "  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.5\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[0] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.10\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[1] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.11\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[2] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.12\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[3] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.13\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[4] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.14\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[5] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.15\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_router_interface_v2.router_interface will be created",
                                "  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {",
                                "      + force_destroy = false",
                                "      + id            = (known after apply)",
                                "      + port_id       = (known after apply)",
                                "      + region        = (known after apply)",
                                "      + router_id     = (known after apply)",
                                "      + subnet_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_router_v2.router will be created",
                                "  + resource \"openstack_networking_router_v2\" \"router\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + distributed             = (known after apply)",
                                "      + enable_snat             = (known after apply)",
                                "      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"",
                                "      + external_qos_policy_id  = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + name                    = \"testbed\"",
                                "      + region                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "",
                                "      + external_fixed_ip (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {",
                                "      + description             = \"ssh\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 22",
                                "      + port_range_min          = 22",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {",
                                "      + description             = \"wireguard\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 51820",
                                "      + port_range_min          = 51820",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {",
                                "      + description             = \"vrrp\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"112\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_management will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"management security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-management\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_node will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"node security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-node\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_subnet_v2.subnet_management will be created",
                                "  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {",
                                "      + all_tags          = (known after apply)",
                                "      + cidr              = \"192.168.16.0/20\"",
                                "      + dns_nameservers   = [",
                                "          + \"8.8.8.8\",",
                                "          + \"9.9.9.9\",",
                                "        ]",
                                "      + enable_dhcp       = true",
                                "      + gateway_ip        = (known after apply)",
                                "      + id                = (known after apply)",
                                "      + ip_version        = 4",
                                "      + ipv6_address_mode = (known after apply)",
                                "      + ipv6_ra_mode      = (known after apply)",
                                "      + name              = \"subnet-testbed-management\"",
                                "      + network_id        = (known after apply)",
                                "      + no_gateway        = false",
                                "      + region            = (known after apply)",
                                "      + service_types     = (known after apply)",
                                "      + tenant_id         = (known after apply)",
                                "",
                                "      + allocation_pool {",
                                "          + end   = \"192.168.31.250\"",
                                "          + start = \"192.168.31.200\"",
                                "        }",
                                "    }",
                                "",
                                "  # terraform_data.image will be created",
                                "  + resource \"terraform_data\" \"image\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "  # terraform_data.image_node will be created",
                                "  + resource \"terraform_data\" \"image_node\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "Plan: 64 to add, 0 to change, 0 to destroy.",
                                "",
                                "Changes to Outputs:",
                                "  + manager_address = (sensitive value)",
                                "  + private_key     = (sensitive value)",
                                "terraform_data.image_node: Creating...",
                                "terraform_data.image_node: Creation complete after 0s [id=e1c2e4bb-a477-452d-9478-bcd164c47803]",
                                "data.openstack_images_image_v2.image_node: Reading...",
                                "openstack_networking_network_v2.net_management: Creating...",
                                "openstack_compute_keypair_v2.key: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creating...",
                                "terraform_data.image: Creating...",
                                "terraform_data.image: Creation complete after 0s [id=d816786d-72d9-acb3-b0bb-a6b8f5fccd60]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creating...",
                                "data.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creating...",
                                "openstack_networking_network_v2.net_management: Creation complete after 1s [id=ab1787c0-a6c8-4860-adff-4d1b12555b92]",
                                "data.openstack_images_image_v2.image: Reading...",
                                "data.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=f69e02e7-d854-4ded-bb8d-51d0e0400336]",
                                "local_file.id_rsa_pub: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=79922d84-0445-4535-976b-32e74e35a748]",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=5295d09e-fddd-4452-8a25-9ba23e2b95ae]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creating...",
                                "local_file.id_rsa_pub: Creation complete after 0s [id=7d94d81cdceac9b3b6345af92bb1be3add46cc11]",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=0a292857-8cd9-4a14-95ba-a5d022f4a90e]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 4s [id=49cd33e4-72cd-4f3f-940d-55c9f0f00a98]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=9db8234e-f6a8-4211-a809-87a509109e78]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 4s [id=5c0bff01-3898-4d25-903e-2ecdf087243c]",
                                "local_sensitive_file.id_rsa: Creating...",
                                "local_sensitive_file.id_rsa: Creation complete after 0s [id=aa1825b291d916f8a2c90b4b6acc8cc92ef06ae8]",
                                "openstack_networking_subnet_v2.subnet_management: Creating...",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 3s [id=1f00cc32-4927-4d99-9c1e-b649b1d1f573]",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=6ad71b9e-76db-4ac5-b372-050f59253056]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 4s [id=4dcaa69c-5414-4861-9f75-cc0da42200e7]",
                                "openstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=aeecf627-ee34-49a3-a604-f1ec5825b8f9]",
                                "openstack_networking_router_v2.router: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=83689365-4423-433a-82c0-63cbcaedfdf8]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=306cfbe9-242f-441d-bc49-37fa1b1f4569]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=5f6d3b65-3844-4001-8889-d6deb3f0644d]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=ffeaaf24-9754-44c8-bb36-eb3a5d2d5315]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=d36bd727-f6fd-4e09-af6c-5d1752a9fb11]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 3s [id=a3f90d2f-3417-480d-afb3-bab2acf5e837]",
                                "openstack_networking_router_v2.router: Creation complete after 4s [id=626bfd75-5d77-4996-991e-e4b06a8e0797]",
                                "openstack_networking_router_interface_v2.router_interface: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creation complete after 0s [id=82a0ffe4-c609-4bad-a240-ed41f4279f28]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...",
                                "openstack_networking_port_v2.manager_port_management: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=58bf471c-d80e-418b-b4d0-705ee3d0ea1c]",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 0s [id=449dd116-1fca-4840-b186-c61271871b7f]",
                                "openstack_networking_port_v2.node_port_management[5]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 0s [id=cd34c19d-4240-4fd7-abcf-9ef72006f683]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=828cb3f4-1eb7-49b3-b1b5-d9c7daab8229]",
                                "openstack_networking_port_v2.node_port_management[4]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=8353e7e2-b281-4ec9-9907-6ebbfc0f389b]",
                                "openstack_networking_port_v2.node_port_management[0]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 1s [id=19dacea6-f8ac-4dce-a10d-e93d240f1da4]",
                                "openstack_networking_port_v2.node_port_management[1]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 1s [id=0490a749-77b9-4f38-91ff-4b1321995ff2]",
                                "openstack_networking_port_v2.node_port_management[2]: Creating...",
                                "openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=21f6ad4e-2daf-49cf-9f3a-83bd124ef2b6]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=3ca4a049-0969-4712-9f12-0f3f489ff07d]",
                                "openstack_networking_port_v2.node_port_management[3]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=f680df32-572e-4bf4-b0da-1405d80f137d]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 1s [id=43d3dec3-2b1f-4793-989e-a338aaadfc07]",
                                "openstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=5133aec3-bcc8-4613-a099-273522a85686]",
                                "openstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=033250be-3730-4e28-8bae-574aa81261e3]",
                                "openstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=b46751e7-e6e1-4843-8018-ec0fe26ce24c]",
                                "openstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=f9c66452-e535-44f2-abde-2653138d6638]",
                                "openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=dc661a00-4fe6-413a-9683-d66479df831d]",
                                "openstack_networking_router_interface_v2.router_interface: Creation complete after 3s [id=e4c1491f-c59c-42b7-8ed3-e8c0ebd3203b]",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creation complete after 2s [id=93e66259-6123-483f-83e0-5ca60df99fb6]",
                                "openstack_compute_instance_v2.node_server[5]: Creating...",
                                "openstack_compute_instance_v2.node_server[3]: Creating...",
                                "openstack_compute_instance_v2.node_server[2]: Creating...",
                                "openstack_compute_instance_v2.node_server[4]: Creating...",
                                "openstack_compute_instance_v2.node_server[0]: Creating...",
                                "openstack_compute_instance_v2.node_server[1]: Creating...",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=1665fef5-3425-4ab0-90ee-a0f1212f09dc]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...",
                                "local_file.MANAGER_ADDRESS: Creating...",
                                "local_file.MANAGER_ADDRESS: Creation complete after 0s [id=01887e5a3def8b233d771971aedf02fc115bae67]",
                                "local_file.inventory: Creating...",
                                "local_file.inventory: Creation complete after 0s [id=b287740adb6f33ee390d5b061cc2099583ba72f7]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=1665fef5-3425-4ab0-90ee-a0f1212f09dc]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Creation complete after 31s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c]",
                                "openstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=71ad69d2-ecf0-4a82-9d7a-7b6f0722f356]",
                                "openstack_compute_instance_v2.node_server[1]: Creation complete after 31s [id=a714327d-623a-4d36-8acf-a2e00af61082]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Creation complete after 41s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b]",
                                "openstack_compute_instance_v2.node_server[4]: Creation complete after 41s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217]",
                                "openstack_compute_instance_v2.node_server[0]: Creation complete after 42s [id=57e5bb4e-daf6-427c-b4df-17c8f7f3e79e]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...",
                                "null_resource.node_semaphore: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...",
                                "null_resource.node_semaphore: Creation complete after 0s [id=7303008649623477253]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...",
                                "openstack_compute_instance_v2.manager_server: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 4s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/5295d09e-fddd-4452-8a25-9ba23e2b95ae]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 4s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/0a292857-8cd9-4a14-95ba-a5d022f4a90e]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 4s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/f69e02e7-d854-4ded-bb8d-51d0e0400336]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/9db8234e-f6a8-4211-a809-87a509109e78]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 4s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/5c0bff01-3898-4d25-903e-2ecdf087243c]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 4s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/1f00cc32-4927-4d99-9c1e-b649b1d1f573]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 10s [id=465cb251-a060-4c51-96e6-b5bbb23ace7b/49cd33e4-72cd-4f3f-940d-55c9f0f00a98]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=4fd6f2b6-51ef-4e2f-bb54-ee6be11bcc4c/79922d84-0445-4535-976b-32e74e35a748]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 10s [id=e7b65721-5a14-41aa-a13c-e8612ce2c217/6ad71b9e-76db-4ac5-b372-050f59253056]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Creation complete after 21s [id=d3b2fcd4-dac4-415f-a170-928342d9466c]",
                                "",
                                "Apply complete! Resources: 64 added, 0 changed, 0 destroyed.",
                                "",
                                "Outputs:",
                                "",
                                "manager_address = <sensitive>",
                                "private_key = <sensitive>"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000009-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:41.255552Z",
                            "start": "2026-01-13T00:02:19.600192Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000009",
                        "name": "Create infrastructure (latest)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_manager_version != \"latest\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:41.824568Z",
                            "start": "2026-01-13T00:03:41.315956Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000a",
                        "name": "Create infrastructure (stable)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTMuMjM0Cg==",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:42.307997Z",
                            "start": "2026-01-13T00:03:41.838703Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000b",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.193.234"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:42.406314Z",
                            "start": "2026-01-13T00:03:42.318755Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000c",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed05/src/github.com/osism/ansible-collection-commons"
                                    ],
                                    "delta": "0:00:00.531083",
                                    "end": "2026-01-13 00:03:43.235257",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed05/src/github.com/osism/ansible-collection-commons\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed05",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "commons",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-01-13 00:03:42.704174",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n[WARNING]: Collection osism.services does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.commons:999.0.0' to '/home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons'\nCreated collection for osism.commons:999.0.0 at /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons\nosism.commons:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.commons:999.0.0' to '/home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons'",
                                        "Created collection for osism.commons:999.0.0 at /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons",
                                        "osism.commons:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed05/src/github.com/osism/ansible-collection-services"
                                    ],
                                    "delta": "0:00:00.814499",
                                    "end": "2026-01-13 00:03:44.326286",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed05/src/github.com/osism/ansible-collection-services\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed05",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "services",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-01-13 00:03:43.511787",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.services does not support Ansible version 2.15.2\n[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/zuul-testbed05/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.services:999.0.0' to '/home/zuul-testbed05/.ansible/collections/ansible_collections/osism/services'",
                                        "Created collection for osism.services:999.0.0 at /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/services",
                                        "osism.services:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:44.392703Z",
                            "start": "2026-01-13T00:03:42.419139Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000d",
                        "name": "Update ansible collections"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 10,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.193.234",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:03:55.452399Z",
                            "start": "2026-01-13T00:03:44.407148Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000e",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-01-13 00:03:55.484762",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-01-13 00:04:55.488227",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:04:55.506365Z",
                            "start": "2026-01-13T00:03:55.464724Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000000f",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:04:57.090119Z",
                            "start": "2026-01-13T00:04:55.576971Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000010",
                        "name": "Fetch manager ssh hostkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "cp /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id* .",
                            "delta": "0:00:00.006816",
                            "end": "2026-01-13 00:04:57.431960",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "cp /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id* .",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000011-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:04:57.425144",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000011-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:04:57.648154Z",
                            "start": "2026-01-13T00:04:57.136639Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000011",
                        "name": "Get ssh keypair from terraform environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:04:57.717517Z",
                            "start": "2026-01-13T00:04:57.666769Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000012",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed05/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                            "delta": "0:05:27.217180",
                            "end": "2026-01-13 00:10:25.307360",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed05/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000013-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:04:58.090180",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Wait for cloud-init to finish] *******************************************\n\nTASK [Check /var/lib/cloud/instance/boot-finished] *****************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Update package cache] ****************************************************\nskipping: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nskipping: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Fail if Ubuntu version is lower than 24.04] ******************************\nskipping: [testbed-manager]\n\nTASK [Fail if Debian version is lower than 12] *********************************\nskipping: [testbed-manager]\n\nTASK [Set APT options on manager] **********************************************\nchanged: [testbed-manager]\n\nTASK [Update APT cache and run dist-upgrade] ***********************************\nchanged: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nchanged: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nchanged: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [Get current user] ********************************************************\nok: [testbed-manager]\n\nTASK [Create venv directory] ***************************************************\nchanged: [testbed-manager]\n\nTASK [Install netaddr in venv] *************************************************\nchanged: [testbed-manager]\n\nTASK [Install ansible-core in venv] ********************************************\nchanged: [testbed-manager]\n\nTASK [Install requests >= 2.32.2] **********************************************\nchanged: [testbed-manager]\n\nTASK [Install docker >= 7.1.0] *************************************************\nchanged: [testbed-manager]\n\nTASK [Create directories in /opt/src] ******************************************\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Sync sources in /opt/src] ************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Create /usr/share/ansible directory] *************************************\nchanged: [testbed-manager]\n\nTASK [Install collections from Ansible galaxy] *********************************\nchanged: [testbed-manager] => (item=ansible.netcommon)\nchanged: [testbed-manager] => (item=ansible.posix)\nchanged: [testbed-manager] => (item=community.docker>=3.10.2)\n\nTASK [Install local collections] ***********************************************\nchanged: [testbed-manager] => (item=ansible-collection-commons)\nchanged: [testbed-manager] => (item=ansible-collection-services)\n\nPLAY [Create operator user] ****************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Create operator group] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Create user] ************************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nchanged: [testbed-manager] => (item=adm)\nchanged: [testbed-manager] => (item=sudo)\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nchanged: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)\nchanged: [testbed-manager] => (item=export LANG=C.UTF-8)\nchanged: [testbed-manager] => (item=export LC_ALL=C.UTF-8)\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set password] ***********************************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Recursively change ownership of /opt/venv] *******************************\nchanged: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Wait for cloud-init to finish] *******************************************",
                                "",
                                "TASK [Check /var/lib/cloud/instance/boot-finished] *****************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Update package cache] ****************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Ubuntu version is lower than 24.04] ******************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Debian version is lower than 12] *********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set APT options on manager] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Update APT cache and run dist-upgrade] ***********************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get current user] ********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Create venv directory] ***************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install netaddr in venv] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install ansible-core in venv] ********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install requests >= 2.32.2] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install docker >= 7.1.0] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create directories in /opt/src] ******************************************",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Sync sources in /opt/src] ************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Create /usr/share/ansible directory] *************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install collections from Ansible galaxy] *********************************",
                                "changed: [testbed-manager] => (item=ansible.netcommon)",
                                "changed: [testbed-manager] => (item=ansible.posix)",
                                "changed: [testbed-manager] => (item=community.docker>=3.10.2)",
                                "",
                                "TASK [Install local collections] ***********************************************",
                                "changed: [testbed-manager] => (item=ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=ansible-collection-services)",
                                "",
                                "PLAY [Create operator user] ****************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "changed: [testbed-manager] => (item=adm)",
                                "changed: [testbed-manager] => (item=sudo)",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LANG=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8)",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Recursively change ownership of /opt/venv] *******************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000013-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:10:25.461337Z",
                            "start": "2026-01-13T00:04:57.758022Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000013",
                        "name": "Run manager part 0"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "It is now already possible to log in to the manager with 'make login'."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:10:25.531495Z",
                            "start": "2026-01-13T00:10:25.478313Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000014",
                        "name": "Point out that the log in on the manager is now possible"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:10:25.594103Z",
                            "start": "2026-01-13T00:10:25.543367Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000015",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed05/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                            "delta": "0:02:21.679023",
                            "end": "2026-01-13 00:12:47.606833",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed05/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000016-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:10:25.927810",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Run manager part 1] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nincluded: /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nskipping: [testbed-manager]\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Update package cache] *************************\nchanged: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Copy SSH public key] *****************************************************\nchanged: [testbed-manager]\n\nTASK [Copy SSH private key] ****************************************************\nchanged: [testbed-manager]\n\nTASK [Create configuration directory] ******************************************\nchanged: [testbed-manager]\n\nTASK [Copy testbed repo] *******************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager]\n\nTASK [Install python requirements in venv] *************************************\nok: [testbed-manager] => (item=Jinja2)\nok: [testbed-manager] => (item=PyYAML)\nok: [testbed-manager] => (item=packaging)\nchanged: [testbed-manager] => (item=python-gilt==1.2.3)\nok: [testbed-manager] => (item=requests>=2.32.2)\nok: [testbed-manager] => (item=docker>=7.1.0)\n\nTASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************\nchanged: [testbed-manager]\n\nTASK [Copy testbed custom CA certificate on CentOS] ****************************\nskipping: [testbed-manager]\n\nTASK [Run update-ca-certificates on Debian/Ubuntu] *****************************\nchanged: [testbed-manager]\n\nTASK [Run update-ca-trust on RedHat] *******************************************\nskipping: [testbed-manager]\n\nTASK [Run manager part 2] ******************************************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nok: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Run manager part 1] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "included: /home/zuul-testbed05/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Copy SSH public key] *****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy SSH private key] ****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create configuration directory] ******************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed repo] *******************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install python requirements in venv] *************************************",
                                "ok: [testbed-manager] => (item=Jinja2)",
                                "ok: [testbed-manager] => (item=PyYAML)",
                                "ok: [testbed-manager] => (item=packaging)",
                                "changed: [testbed-manager] => (item=python-gilt==1.2.3)",
                                "ok: [testbed-manager] => (item=requests>=2.32.2)",
                                "ok: [testbed-manager] => (item=docker>=7.1.0)",
                                "",
                                "TASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed custom CA certificate on CentOS] ****************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-certificates on Debian/Ubuntu] *****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-trust on RedHat] *******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run manager part 2] ******************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000016-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:12:47.750706Z",
                            "start": "2026-01-13T00:10:25.630318Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000016",
                        "name": "Run manager part 1 + 2"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "sudo",
                                "shutdown",
                                "-r",
                                "now"
                            ],
                            "delta": "0:00:00.963978",
                            "end": "2026-01-13 00:12:49.043662",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 sudo shutdown -r now",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000017-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:12:48.079684",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-000000000017-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:12:49.312833Z",
                            "start": "2026-01-13T00:12:47.800690Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000017",
                        "name": "Reboot manager"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 18,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.193.234",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:13:08.249650Z",
                            "start": "2026-01-13T00:12:49.328551Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000018",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-01-13 00:13:08.285335",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-01-13 00:14:08.288071",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:14:08.307242Z",
                            "start": "2026-01-13T00:13:08.258397Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000019",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "/opt/configuration/scripts/deploy-manager.sh"
                            ],
                            "delta": "0:24:20.888740",
                            "end": "2026-01-13 00:38:29.913666",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 /opt/configuration/scripts/deploy-manager.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001a-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:14:09.024926",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY MANAGER\n\n+ set -e\n+ echo\n+ echo '# DEPLOY MANAGER'\n+ echo\n+ cat /opt/manager-vars.sh\nexport NUMBER_OF_NODES=6\n\nexport CEPH_VERSION=reef\nexport CONFIGURATION_VERSION=main\nexport MANAGER_VERSION=latest\nexport OPENSTACK_VERSION=2024.2\n\nexport ARA=false\nexport DEPLOY_MODE=manager\nexport TEMPEST=true\nexport IS_ZUUL=true\n\nexport MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\nexport EXTERNAL_API=false\n\nexport IMAGE_USER=ubuntu\nexport IMAGE_NODE_USER=ubuntu\n\nexport CEPH_STACK=ceph-ansible\n\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver\n+ docker version\nClient: Docker Engine - Community\n Version:           27.5.1\n API version:       1.47\n Go version:        go1.22.11\n Git commit:        9f9e405\n Built:             Wed Jan 22 13:41:48 2025\n OS/Arch:           linux/amd64\n Context:           default\n\nServer: Docker Engine - Community\n Engine:\n  Version:          27.5.1\n  API version:      1.47 (minimum version 1.24)\n  Go version:       go1.22.11\n  Git commit:       4c9b3b0\n  Built:            Wed Jan 22 13:41:48 2025\n  OS/Arch:          linux/amd64\n  Experimental:     false\n containerd:\n  Version:          v2.2.1\n  GitCommit:        dea7da592f5d1d2b7755e3a161be07f43fad8f75\n runc:\n  Version:          1.3.4\n  GitCommit:        v1.3.4-0-gd6d73eb8\n docker-init:\n  Version:          0.19.0\n  GitCommit:        de40ad0\n+ sh -c /opt/configuration/scripts/deploy/000-manager.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-ceph-version.sh reef\n+ set -e\n+ VERSION=reef\n++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n ceph_version: reef ]]\n+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml\n+ /opt/configuration/scripts/set-openstack-version.sh 2024.2\n+ set -e\n+ VERSION=2024.2\n++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n openstack_version: 2024.2 ]]\n+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ ceph-ansible == \\r\\o\\o\\k ]]\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ echo 'enable_osism_kubernetes: true'\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n++ semver 2024.2 2025.1\n+ [[ -1 -ge 0 ]]\n+ /opt/configuration/scripts/enable-resource-nodes.sh\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml\n\nPLAY [Copy custom facts] *******************************************************\n\nTASK [Create custom facts directory] *******************************************\nok: [testbed-manager]\n\nTASK [Copy fact files] *********************************************************\nchanged: [testbed-manager]\n\nPLAY [Before the deployment of the manager] ************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get /opt/manager-vars.sh] ************************************************\nok: [testbed-manager]\n\nTASK [Add ara_server_mariadb_volume_type parameter] ****************************\nchanged: [testbed-manager]\n\nTASK [Add netbox_enable parameter] *********************************************\nskipping: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Use insecure glance configuration] ***************************************\nskipping: [testbed-manager]\n\nTASK [Check if /etc/OTC_region exist] ******************************************\nok: [testbed-manager]\n\nTASK [Add nova_compute_virt_type parameter] ************************************\nskipping: [testbed-manager]\n\nPLAY [Apply role traefik] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply traefik role] ******************************************************\nincluded: osism.services.traefik for testbed-manager\n\nTASK [osism.services.traefik : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager\n\nTASK [osism.services.traefik : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/traefik)\nchanged: [testbed-manager] => (item=/opt/traefik/certificates)\nchanged: [testbed-manager] => (item=/opt/traefik/configuration)\n\nTASK [osism.services.traefik : Copy configuration files] ***********************\nchanged: [testbed-manager] => (item=traefik.yml)\nchanged: [testbed-manager] => (item=traefik.env)\nchanged: [testbed-manager] => (item=certificates.yml)\n\nTASK [osism.services.traefik : Copy certificate cert files] ********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy certificate key files] *********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy dynamic configuration] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.traefik : Remove dynamic configuration] *******************\nok: [testbed-manager]\n\nTASK [osism.services.traefik : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager\n\nTASK [osism.services.traefik : Create traefik external network] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Manage traefik service] *************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.traefik : Restart traefik service] *************\nskipping: [testbed-manager]\n\nPLAY [Deploy manager service] **************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply manager role] ******************************************************\nincluded: osism.services.manager for testbed-manager\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/ansible)\nchanged: [testbed-manager] => (item=/opt/archive)\nchanged: [testbed-manager] => (item=/opt/manager/configuration)\nchanged: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nchanged: [testbed-manager] => (item=/opt/manager/secrets)\nchanged: [testbed-manager] => (item=/opt/ansible/secrets)\nchanged: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nchanged: [testbed-manager] => (item=ara)\nchanged: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nchanged: [testbed-manager] => (item=conductor)\nchanged: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nchanged: [testbed-manager] => (item=osism)\nchanged: [testbed-manager] => (item=osism-update-docker)\nchanged: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nchanged: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nok: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nskipping: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=13   rescued=0    ignored=0   \n\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   \nmanager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago        Up About a minute (healthy)   192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago        Up About a minute (healthy)   8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago        Up About a minute (healthy)   \nmanager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago        Up About a minute (healthy)   \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 57 seconds (healthy)       \nmanager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago        Up About a minute (healthy)   \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago        Up About a minute (healthy)   3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago        Up About a minute (healthy)   \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago        Up About a minute (healthy)   6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   \nosism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago        Up About a minute             192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   \nosismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago        Up About a minute (healthy)   \n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg\n+ osism apply resolvconf -l testbed-manager\n\u001b[32m2026-01-13 00:18:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d37815b-b49d-49a6-90f4-18332c355b3c (resolvconf) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:18:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d37815b-b49d-49a6-90f4-18332c355b3c (resolvconf) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role resolvconf] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 13 January 2026  00:18:55 +0000 (0:00:00.144)       0:00:00.144 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nTuesday 13 January 2026  00:18:58 +0000 (0:00:03.602)       0:00:03.746 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nTuesday 13 January 2026  00:18:58 +0000 (0:00:00.067)       0:00:03.814 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nTuesday 13 January 2026  00:18:58 +0000 (0:00:00.089)       0:00:03.904 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nTuesday 13 January 2026  00:18:58 +0000 (0:00:00.073)       0:00:03.978 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nTuesday 13 January 2026  00:19:00 +0000 (0:00:01.066)       0:00:05.045 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nTuesday 13 January 2026  00:19:00 +0000 (0:00:00.063)       0:00:05.108 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nTuesday 13 January 2026  00:19:00 +0000 (0:00:00.476)       0:00:05.585 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nTuesday 13 January 2026  00:19:00 +0000 (0:00:00.075)       0:00:05.660 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nTuesday 13 January 2026  00:19:01 +0000 (0:00:00.529)       0:00:06.190 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nTuesday 13 January 2026  00:19:02 +0000 (0:00:01.060)       0:00:07.250 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nTuesday 13 January 2026  00:19:03 +0000 (0:00:00.951)       0:00:08.201 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nTuesday 13 January 2026  00:19:03 +0000 (0:00:00.086)       0:00:08.287 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:19:04 +0000 (0:00:01.139)       0:00:09.427 ******* \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.60s\nosism.commons.resolvconf : Restart systemd-resolved service ------------- 1.14s\nosism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.07s\nosism.commons.resolvconf : Copy configuration files --------------------- 1.06s\nosism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.95s\nosism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.53s\nosism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.48s\nosism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s\nosism.commons.resolvconf : Include distribution specific configuration tasks --- 0.09s\nosism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.08s\nosism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s\nosism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s\nosism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s\n+ osism apply sshconfig\n\u001b[32m2026-01-13 00:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 089338ef-ce47-42f2-90d9-dd2487ae976e (sshconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 089338ef-ce47-42f2-90d9-dd2487ae976e (sshconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role sshconfig] ****************************************************\n\nTASK [osism.commons.sshconfig : Get home directory of operator user] ***********\nTuesday 13 January 2026  00:19:20 +0000 (0:00:00.155)       0:00:00.155 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************\nTuesday 13 January 2026  00:19:21 +0000 (0:00:00.564)       0:00:00.719 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure config for each host exist] *************\nTuesday 13 January 2026  00:19:21 +0000 (0:00:00.400)       0:00:01.119 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\nTASK [osism.commons.sshconfig : Add extra config] ******************************\nTuesday 13 January 2026  00:19:27 +0000 (0:00:05.136)       0:00:06.256 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Assemble ssh config] ***************************\nTuesday 13 January 2026  00:19:27 +0000 (0:00:00.081)       0:00:06.337 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:19:27 +0000 (0:00:00.575)       0:00:06.912 ******* \n=============================================================================== \nosism.commons.sshconfig : Ensure config for each host exist ------------- 5.14s\nosism.commons.sshconfig : Assemble ssh config --------------------------- 0.58s\nosism.commons.sshconfig : Get home directory of operator user ----------- 0.56s\nosism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.40s\nosism.commons.sshconfig : Add extra config ------------------------------ 0.08s\n+ osism apply known-hosts\n\u001b[32m2026-01-13 00:19:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30ba9bc7-417f-4a1b-9d72-ca014fae44b4 (known-hosts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:19:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 30ba9bc7-417f-4a1b-9d72-ca014fae44b4 (known-hosts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role known_hosts] **************************************************\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***\nTuesday 13 January 2026  00:19:44 +0000 (0:00:00.122)       0:00:00.122 ******* \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***\nTuesday 13 January 2026  00:19:50 +0000 (0:00:05.749)       0:00:05.872 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:50 +0000 (0:00:00.161)       0:00:06.033 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEwdqUB7HN4VL2b39E+wmGAVzyJqO2pg6Y5v1/wgoBcJnDKVoLinltTeLID6TqK0rNRIrSeCdJ7ghn9+OntawF8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDhG4+2lbFna6pQJ31DRXMxGyaPNBhYP+SgYYY5Vj+SwWIJllHlv7shVTlowhDNWJH4AM5sgwNoKcSveeq58h3JoZrMC54iWMBHf7+7MpWa/kdQq9he91PMOGXYjjoaF4+l0qRkMDWTLgNz8E1qQEcH2xX89F+GlZF/3Kb8EaAnLe2ckhBlt+dydNlJPgNWT5hjgHgHCeVYe2KKhlXogk4CFVQr5YwnRO0O0vOONQ2Ryefm146Wyld/tV/22YAA7LItrYkJNGl14yYQy38A5NQUnMgHzCMMFzrPhz/IVuy8TUKyds1DrVjv1ZQjCLM1V2Xn5ENanCKlDLy1Sg774GRHwyOvoER4GADruHMhUn7DFJQo3a6vr7rUFq8gtvV6L54EFVDiuO0XqztSa4uSirBPZv0G/jM9fLDt850YS+jN2nFQATOYQvt5ADRFqUOs1MUfP02mCECk4SOakrIPRINL/gIUrFAbah16IovwWreAg4EuzwOSi0mG5l1P4wrRztk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIG1wld3KLY273VgRhq3DY4Y5VreDQGtCBE/2+XB2y63y)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:51 +0000 (0:00:01.143)       0:00:07.177 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAII3cBlxJ+Nj+vcX33DDsiEGmFzhWhUplm9v08bHaGxHd)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDMNy4Ll3tlu7tXNUF3crkWBEoB1g1PQzruThglC08FOVDD6ejeWmgWy/vre8E+ErGv22sgxBVVepNWHM6ektcqb92n6EoG348bB9Zc6/dZZEEDIMvkuDQujy4Q+enMJKou3EIC8IkKUz7Qwp5nIKTTm0mOLXkmrXMYyA5TfdPaYwfncOIOmbydM05j6gD4KbIxQabeevh29UxN2KeEQTBh0Aaiw/CJTPW/sObRcdr6PiyCWPiUVAZeFSzS2ESfTVBwO0CLhtc5+lG3Z8jS0Asnu+hH6BLqzoU2I07edidZM3QiAtugKAbmdzf7f1+a88PgwxBGqT4AeBrwN4BoNVQBWYcP3fwnASx9koirDoWHqT5As8SUc2Zi/AcNUsrVpgwhxpafm4SkF+gvu3v1gFxhJ3+2bFbdrKLlDwxMM0vOvqHYAF32q9PM5vY9ZsqaDY+R2kYw65yed75mlfT+z11I7mTLiJegVMIayC9Ms6ULIyqKkpuE35qJx+zeUHS0D/M=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCANVsfM+rupUfrqcS59qXGwJqieAokpdxEwmlCCPh+A0evjKfvhj+QDlIGxgTNnid4YWZD3owkOldDVz1czbl4=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:52 +0000 (0:00:01.111)       0:00:08.288 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCs4yVojS8g9mYfYUGVbXOe9f9Q4SCvWhW3wolwzYPHIv9Xbbq3IpBpMmRarHfb4EbZ7MLtSbcxXJIWOJ5uT/v1eyT5T+8N2wVHQcZDGLxEAXU6P/wxNX1zzO9k+aHpUchtOIy01JN2dXZs5Sy5k1ZSuceJa3W3bMEdGGZ2875Lx5KDP/ivLciez4rUjtCdI0zGw/yIFfZDmiCMNBf1J4CJkuRRtEPnGhvch3AwM/2ma2fJOCFqDynv4wsIAH+5wXXeGKmtJhApo2jdhDpXlx8dDw8iE638+4gHpBb7t6EnvtfAUbp1BPdruiocm85F7TrtzepT0F8F7bCHuCCsDhZ+7Lq5VyD1GBXQTy9T4txIbj1cfYd7pP1ztsCNGtsfs2Te2qzihxL09FSWaKXe3oH0kyyjs6oAGJEPUYIa14l0Nhwf/F/Y39NcTJcpKQO5cVItDmypoSsRW/FmwWSfj4qLgjvG6laR+WwQtQ2BHQKU5OEHMaUO4nPwpOaCldDAcHk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIIb9uQuSUhF1yMMy/r7j8CSdaGWoRrCFZZtsbBuHk1lnHQp2c8Q2i0dbJwrz2NEaAHO2XIpenm+NecOa6v3p2U=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJDk/kf1I0kvvlU6paCuuF/kJCN7s38Ieq6GukgVT+0z)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:53 +0000 (0:00:01.034)       0:00:09.323 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBBT3kQanK+MPgr+KonMfs6XdQsNGu/kU8ttz39IgxJTK4uiJwzFrECUpR4sjnyGb7etmLer6xO0aafH+oD8Jc4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDSzAl+AST4boAzW56Ko76Dtd1M0bcRU4tyJFCjJH0MXVG7lRT6+RfPSA0kGrSjafm6IWd4oVyiHfYmxLxlw9qbyURve7RwylY9CnXoklCsk/mlNJ+OA6ev3zoe/1+oRcgf8zLlUgytux5eyVs0gQL0SqQ10VPq0h0zIJplqPKKhlbQQAxps722qV7ak0MWY6iet9q+AQvrnQUK9dUyhHjP12RIQaUUNtn7JJQT/wUNAk+1nNKSTrb9mnQd5Rf9ak/fxAzBHjTctwf8f5p5GURoEAwRbUikWILvHlP8adxBdS0LfmqhS0W8WyGSb70yReIKzWpIxbjf9nkEwaAYrAXsvwKSGZXz6wGtri/OvXaHpXd86ysh7zku38aMFCpkZNUPaok6j8AQ7Wwk/HgE6BM6DlVjjttr7tnmuiia7nH7El2/x06QeaTHBHD9LAyvSbfaLGr9RldoEHv3GrSeVgd6zmlsWAuJH+dMbbbii4vfL+snp5Uv+ZJYpM34r2S7Fp8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKKhUM0juZlhzLj/K6dKqkuv5KvYz7H8RaW5TIFwpjIo)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:54 +0000 (0:00:01.023)       0:00:10.346 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDNRkg+yOke3Ddbei3PVLXSl/gJWdLCOPL0+k2kVStf6W6HnoSVqAA5dQQ3nXjW8qov3yO+0olH+j2KXOlaK3tY=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiIV9KgZPNLypT6BgMveQMLb+XP2pvowAeWbYSVy5SRDT4C/PTEvxgtPE0/a2Uq1ynUXQr9qfIMI1ISnuQ2P189RXwLdzXsuECbWwQRc+iTH+7NZOnsONukNVYPuQw772ZM+l/r6CdNv0Z0dYcm9DO7MQGAWSrEZoMCeTfvACeVD+SxZNioE13o0N+riGgnwJppHHMjh5ILfJ7eu9yfKGgqOij82UBLwmWSliga7vkTObjQ5UyfGeju1sf7wsx1akOGSgbFTFRUvaQPUAJJ5G862lRN5n+iXZvlytzFDs8UFgVnRDPSRD+96KQ9c2An7hpBMe1EJHbaaHWB4btrOcRrZY2dzbN3rrVKCTuf2v/Sju6B/hexQ582byuG6WUf7Q5mw2HV0yRct3CpB8rCwKy1IKdtrjncIQ47G5XAfpDWGufIG4h14ZgdO4Gb9zp8xRi11VauLn1YL2YuTVWqgbrjeCXA7gmsrLGpa2dOV1n/ByXe53b4DQr1gq0XufLzak=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILYwppQSAKDyEbxkq0PEPVz0XhqRpbAsz1hlj5fhJOjq)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:55 +0000 (0:00:00.974)       0:00:11.320 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBiN8QzGQv8ZHPfLAudwKKlhtdteTM1C6Cw2AUpga87mGiT8xUv81/alwHnXJq3Ey3e9/R5sm9J1abxgzuI1J8M7owK/mN4T2LlXSJHlagJNjQvfKlmxZz1l+kppPMKfXcaLTwzp9gQUsMjTj4UGuO/PM4ldgmbE/NhWO6afrVT/tJnFN+UM2UgJ5BCR2N7rtPBPmI00CQGWr/WPcjDIT9aFNrZR9wr4xJwhSDePHSUAN8QjR3AEfr9ihcSBM05ESY9JhmHgFHiWyYsoCclIMclmBl+c7Bb2Sb0PlBoEH6kaM+vWS68L4i7A1EHJ/muJco6MYUofRimAYxW2+fS+Ehb6KA9Cse2F5u/b5rRcD/ZOoRgnvbl/e4M7EYCUrzOE2exqKl1Bl6hyJz4gmBejh7/Hoh0vIKqRryeuojSyAr76CegIJeSVMXHuF2l3V6qZ28UaPelyYqGbyQZdSG4S5uJOt+Cwp0neF9Egk2Sr/wvzmo0CMhzcekdV1l0uwp8CM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOpm2DX+iuhXBHOqADVnwtARbnSmUmbxHxRfUznOX5jn2moh3ojj7q8FUsWlwITOW4zsjw7yBh1DDEJH7LBlkQ4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBRKL3VzcKtOBW81QKJd742xWys9tDvetvUB2cavjZrf)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:19:56 +0000 (0:00:00.973)       0:00:12.293 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC2PZmaIQFk5fXdGfNQ+ybya9B5NctgaD8n6aoYA9hO92qMhg4UMz3L7ulQZarThvALWLN043WlfxapsN/zRnL6sK/yIXN+Em7Opi+v5NQCebcQB7xR0NB7dKLwW2ocTgjkNn8SkZ5scnUUh56VdCz+tpXorPuAGfbGPBuHjF35oiCOezymbJ4aeWj+wP/JeyE5XLWSS79QoosNLYfj2Yunul8lbPJPdSjXMXOp/DLuj/u/RM5jN0O5JU1digqgRnnQGu6TGu4le4RaDc9XozNlkrt6sN8fLwMC3oeIex0CWE1HoLS41VA6PHUlSabJ1xksQ0VpmontWHSIBuFvNAKG8gaFxIVlolm8T91B8wKx4F9hcZLgGvkhoAzancJ5itgN7xFopLQLa+D+GaFlta7DXGTTz4LllEb52vWB6U/4MIsaOJ3HDdss7MTR3vlK4mJgA0IZ8OPeitHhpeeycZPT8lumHDpN1f1W8d91App9YY4k4ZED5u052Tmh0fEV+RM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPc5npEgejWX+Bx8E+UvNX2sIZ3DqVmvOFlDMHrWgyf787GSnJk71oQxAe6VCEQGIy0fCbdGZCnpoA9V/nv1UtQ=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGOl+o72g5WDnUdyk5PWe2VS1qTxhFD/4XNOHzgBM+4b)\u001b[0m\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***\nTuesday 13 January 2026  00:19:58 +0000 (0:00:02.030)       0:00:14.324 ******* \n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***\nTuesday 13 January 2026  00:20:03 +0000 (0:00:05.221)       0:00:19.546 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:03 +0000 (0:00:00.166)       0:00:19.713 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEwdqUB7HN4VL2b39E+wmGAVzyJqO2pg6Y5v1/wgoBcJnDKVoLinltTeLID6TqK0rNRIrSeCdJ7ghn9+OntawF8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDhG4+2lbFna6pQJ31DRXMxGyaPNBhYP+SgYYY5Vj+SwWIJllHlv7shVTlowhDNWJH4AM5sgwNoKcSveeq58h3JoZrMC54iWMBHf7+7MpWa/kdQq9he91PMOGXYjjoaF4+l0qRkMDWTLgNz8E1qQEcH2xX89F+GlZF/3Kb8EaAnLe2ckhBlt+dydNlJPgNWT5hjgHgHCeVYe2KKhlXogk4CFVQr5YwnRO0O0vOONQ2Ryefm146Wyld/tV/22YAA7LItrYkJNGl14yYQy38A5NQUnMgHzCMMFzrPhz/IVuy8TUKyds1DrVjv1ZQjCLM1V2Xn5ENanCKlDLy1Sg774GRHwyOvoER4GADruHMhUn7DFJQo3a6vr7rUFq8gtvV6L54EFVDiuO0XqztSa4uSirBPZv0G/jM9fLDt850YS+jN2nFQATOYQvt5ADRFqUOs1MUfP02mCECk4SOakrIPRINL/gIUrFAbah16IovwWreAg4EuzwOSi0mG5l1P4wrRztk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIG1wld3KLY273VgRhq3DY4Y5VreDQGtCBE/2+XB2y63y)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:04 +0000 (0:00:01.018)       0:00:20.731 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAII3cBlxJ+Nj+vcX33DDsiEGmFzhWhUplm9v08bHaGxHd)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDMNy4Ll3tlu7tXNUF3crkWBEoB1g1PQzruThglC08FOVDD6ejeWmgWy/vre8E+ErGv22sgxBVVepNWHM6ektcqb92n6EoG348bB9Zc6/dZZEEDIMvkuDQujy4Q+enMJKou3EIC8IkKUz7Qwp5nIKTTm0mOLXkmrXMYyA5TfdPaYwfncOIOmbydM05j6gD4KbIxQabeevh29UxN2KeEQTBh0Aaiw/CJTPW/sObRcdr6PiyCWPiUVAZeFSzS2ESfTVBwO0CLhtc5+lG3Z8jS0Asnu+hH6BLqzoU2I07edidZM3QiAtugKAbmdzf7f1+a88PgwxBGqT4AeBrwN4BoNVQBWYcP3fwnASx9koirDoWHqT5As8SUc2Zi/AcNUsrVpgwhxpafm4SkF+gvu3v1gFxhJ3+2bFbdrKLlDwxMM0vOvqHYAF32q9PM5vY9ZsqaDY+R2kYw65yed75mlfT+z11I7mTLiJegVMIayC9Ms6ULIyqKkpuE35qJx+zeUHS0D/M=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCANVsfM+rupUfrqcS59qXGwJqieAokpdxEwmlCCPh+A0evjKfvhj+QDlIGxgTNnid4YWZD3owkOldDVz1czbl4=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:05 +0000 (0:00:01.026)       0:00:21.758 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCs4yVojS8g9mYfYUGVbXOe9f9Q4SCvWhW3wolwzYPHIv9Xbbq3IpBpMmRarHfb4EbZ7MLtSbcxXJIWOJ5uT/v1eyT5T+8N2wVHQcZDGLxEAXU6P/wxNX1zzO9k+aHpUchtOIy01JN2dXZs5Sy5k1ZSuceJa3W3bMEdGGZ2875Lx5KDP/ivLciez4rUjtCdI0zGw/yIFfZDmiCMNBf1J4CJkuRRtEPnGhvch3AwM/2ma2fJOCFqDynv4wsIAH+5wXXeGKmtJhApo2jdhDpXlx8dDw8iE638+4gHpBb7t6EnvtfAUbp1BPdruiocm85F7TrtzepT0F8F7bCHuCCsDhZ+7Lq5VyD1GBXQTy9T4txIbj1cfYd7pP1ztsCNGtsfs2Te2qzihxL09FSWaKXe3oH0kyyjs6oAGJEPUYIa14l0Nhwf/F/Y39NcTJcpKQO5cVItDmypoSsRW/FmwWSfj4qLgjvG6laR+WwQtQ2BHQKU5OEHMaUO4nPwpOaCldDAcHk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIIb9uQuSUhF1yMMy/r7j8CSdaGWoRrCFZZtsbBuHk1lnHQp2c8Q2i0dbJwrz2NEaAHO2XIpenm+NecOa6v3p2U=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJDk/kf1I0kvvlU6paCuuF/kJCN7s38Ieq6GukgVT+0z)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:07 +0000 (0:00:01.050)       0:00:22.808 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBBT3kQanK+MPgr+KonMfs6XdQsNGu/kU8ttz39IgxJTK4uiJwzFrECUpR4sjnyGb7etmLer6xO0aafH+oD8Jc4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDSzAl+AST4boAzW56Ko76Dtd1M0bcRU4tyJFCjJH0MXVG7lRT6+RfPSA0kGrSjafm6IWd4oVyiHfYmxLxlw9qbyURve7RwylY9CnXoklCsk/mlNJ+OA6ev3zoe/1+oRcgf8zLlUgytux5eyVs0gQL0SqQ10VPq0h0zIJplqPKKhlbQQAxps722qV7ak0MWY6iet9q+AQvrnQUK9dUyhHjP12RIQaUUNtn7JJQT/wUNAk+1nNKSTrb9mnQd5Rf9ak/fxAzBHjTctwf8f5p5GURoEAwRbUikWILvHlP8adxBdS0LfmqhS0W8WyGSb70yReIKzWpIxbjf9nkEwaAYrAXsvwKSGZXz6wGtri/OvXaHpXd86ysh7zku38aMFCpkZNUPaok6j8AQ7Wwk/HgE6BM6DlVjjttr7tnmuiia7nH7El2/x06QeaTHBHD9LAyvSbfaLGr9RldoEHv3GrSeVgd6zmlsWAuJH+dMbbbii4vfL+snp5Uv+ZJYpM34r2S7Fp8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKKhUM0juZlhzLj/K6dKqkuv5KvYz7H8RaW5TIFwpjIo)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:08 +0000 (0:00:01.046)       0:00:23.855 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDNRkg+yOke3Ddbei3PVLXSl/gJWdLCOPL0+k2kVStf6W6HnoSVqAA5dQQ3nXjW8qov3yO+0olH+j2KXOlaK3tY=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiIV9KgZPNLypT6BgMveQMLb+XP2pvowAeWbYSVy5SRDT4C/PTEvxgtPE0/a2Uq1ynUXQr9qfIMI1ISnuQ2P189RXwLdzXsuECbWwQRc+iTH+7NZOnsONukNVYPuQw772ZM+l/r6CdNv0Z0dYcm9DO7MQGAWSrEZoMCeTfvACeVD+SxZNioE13o0N+riGgnwJppHHMjh5ILfJ7eu9yfKGgqOij82UBLwmWSliga7vkTObjQ5UyfGeju1sf7wsx1akOGSgbFTFRUvaQPUAJJ5G862lRN5n+iXZvlytzFDs8UFgVnRDPSRD+96KQ9c2An7hpBMe1EJHbaaHWB4btrOcRrZY2dzbN3rrVKCTuf2v/Sju6B/hexQ582byuG6WUf7Q5mw2HV0yRct3CpB8rCwKy1IKdtrjncIQ47G5XAfpDWGufIG4h14ZgdO4Gb9zp8xRi11VauLn1YL2YuTVWqgbrjeCXA7gmsrLGpa2dOV1n/ByXe53b4DQr1gq0XufLzak=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILYwppQSAKDyEbxkq0PEPVz0XhqRpbAsz1hlj5fhJOjq)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:09 +0000 (0:00:01.059)       0:00:24.915 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBRKL3VzcKtOBW81QKJd742xWys9tDvetvUB2cavjZrf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBiN8QzGQv8ZHPfLAudwKKlhtdteTM1C6Cw2AUpga87mGiT8xUv81/alwHnXJq3Ey3e9/R5sm9J1abxgzuI1J8M7owK/mN4T2LlXSJHlagJNjQvfKlmxZz1l+kppPMKfXcaLTwzp9gQUsMjTj4UGuO/PM4ldgmbE/NhWO6afrVT/tJnFN+UM2UgJ5BCR2N7rtPBPmI00CQGWr/WPcjDIT9aFNrZR9wr4xJwhSDePHSUAN8QjR3AEfr9ihcSBM05ESY9JhmHgFHiWyYsoCclIMclmBl+c7Bb2Sb0PlBoEH6kaM+vWS68L4i7A1EHJ/muJco6MYUofRimAYxW2+fS+Ehb6KA9Cse2F5u/b5rRcD/ZOoRgnvbl/e4M7EYCUrzOE2exqKl1Bl6hyJz4gmBejh7/Hoh0vIKqRryeuojSyAr76CegIJeSVMXHuF2l3V6qZ28UaPelyYqGbyQZdSG4S5uJOt+Cwp0neF9Egk2Sr/wvzmo0CMhzcekdV1l0uwp8CM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOpm2DX+iuhXBHOqADVnwtARbnSmUmbxHxRfUznOX5jn2moh3ojj7q8FUsWlwITOW4zsjw7yBh1DDEJH7LBlkQ4=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 13 January 2026  00:20:10 +0000 (0:00:01.040)       0:00:25.955 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC2PZmaIQFk5fXdGfNQ+ybya9B5NctgaD8n6aoYA9hO92qMhg4UMz3L7ulQZarThvALWLN043WlfxapsN/zRnL6sK/yIXN+Em7Opi+v5NQCebcQB7xR0NB7dKLwW2ocTgjkNn8SkZ5scnUUh56VdCz+tpXorPuAGfbGPBuHjF35oiCOezymbJ4aeWj+wP/JeyE5XLWSS79QoosNLYfj2Yunul8lbPJPdSjXMXOp/DLuj/u/RM5jN0O5JU1digqgRnnQGu6TGu4le4RaDc9XozNlkrt6sN8fLwMC3oeIex0CWE1HoLS41VA6PHUlSabJ1xksQ0VpmontWHSIBuFvNAKG8gaFxIVlolm8T91B8wKx4F9hcZLgGvkhoAzancJ5itgN7xFopLQLa+D+GaFlta7DXGTTz4LllEb52vWB6U/4MIsaOJ3HDdss7MTR3vlK4mJgA0IZ8OPeitHhpeeycZPT8lumHDpN1f1W8d91App9YY4k4ZED5u052Tmh0fEV+RM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGOl+o72g5WDnUdyk5PWe2VS1qTxhFD/4XNOHzgBM+4b)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPc5npEgejWX+Bx8E+UvNX2sIZ3DqVmvOFlDMHrWgyf787GSnJk71oQxAe6VCEQGIy0fCbdGZCnpoA9V/nv1UtQ=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write static known_hosts entries] ************\nTuesday 13 January 2026  00:20:11 +0000 (0:00:01.083)       0:00:27.038 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Write extra known_hosts entries] *************\nTuesday 13 January 2026  00:20:11 +0000 (0:00:00.156)       0:00:27.195 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Delete known_hosts entries] ******************\nTuesday 13 January 2026  00:20:11 +0000 (0:00:00.066)       0:00:27.261 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Set file permissions] ************************\nTuesday 13 January 2026  00:20:11 +0000 (0:00:00.055)       0:00:27.317 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:20:12 +0000 (0:00:00.703)       0:00:28.020 ******* \n=============================================================================== \nosism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.75s\nosism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.22s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 2.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.14s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.11s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.08s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.06s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s\nosism.commons.known_hosts : Set file permissions ------------------------ 0.70s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s\nosism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s\n+ osism apply squid\n\u001b[32m2026-01-13 00:20:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e69aa7d1-79c0-4f37-a147-7b4d8d64571b (squid) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:20:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e69aa7d1-79c0-4f37-a147-7b4d8d64571b (squid) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role squid] ********************************************************\n\nTASK [osism.services.squid : Include install tasks] ****************************\nTuesday 13 January 2026  00:20:28 +0000 (0:00:00.174)       0:00:00.174 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.squid : Install required packages] ************************\nTuesday 13 January 2026  00:20:28 +0000 (0:00:00.085)       0:00:00.260 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Create required directories] **********************\nTuesday 13 January 2026  00:20:30 +0000 (0:00:01.433)       0:00:01.694 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m\n\nTASK [osism.services.squid : Copy squid configuration files] *******************\nTuesday 13 January 2026  00:20:31 +0000 (0:00:01.121)       0:00:02.816 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m\n\nTASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***\nTuesday 13 January 2026  00:20:32 +0000 (0:00:01.093)       0:00:03.910 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Copy docker-compose.yml file] *********************\nTuesday 13 January 2026  00:20:32 +0000 (0:00:00.316)       0:00:04.226 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Manage squid service] *****************************\nTuesday 13 January 2026  00:20:33 +0000 (0:00:00.901)       0:00:05.128 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Restart squid service] *****************\nTuesday 13 January 2026  00:21:04 +0000 (0:00:30.388)       0:00:35.516 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******\nTuesday 13 January 2026  00:21:19 +0000 (0:00:15.719)       0:00:51.236 ******* \nPausing for 60 seconds\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***\nTuesday 13 January 2026  00:22:20 +0000 (0:01:00.081)       0:01:51.317 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****\nTuesday 13 January 2026  00:22:20 +0000 (0:00:00.067)       0:01:51.385 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:22:20 +0000 (0:00:00.568)       0:01:51.954 ******* \n=============================================================================== \nosism.services.squid : Wait for squid service to start ----------------- 60.08s\nosism.services.squid : Manage squid service ---------------------------- 30.39s\nosism.services.squid : Restart squid service --------------------------- 15.72s\nosism.services.squid : Install required packages ------------------------ 1.43s\nosism.services.squid : Create required directories ---------------------- 1.12s\nosism.services.squid : Copy squid configuration files ------------------- 1.09s\nosism.services.squid : Copy docker-compose.yml file --------------------- 0.90s\nosism.services.squid : Wait for an healthy squid service ---------------- 0.57s\nosism.services.squid : Remove osism_allow_list.conf configuration file --- 0.32s\nosism.services.squid : Include install tasks ---------------------------- 0.09s\nosism.services.squid : Register that squid service was restarted -------- 0.07s\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla\n+ set -e\n+ NAMESPACE=kolla\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n++ semver latest 9.0.0\n+ [[ -1 -lt 0 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ osism apply operator -u ubuntu -l testbed-nodes\n\u001b[32m2026-01-13 00:22:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06e949a0-1f23-46bf-b910-fb634889464f (operator) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:22:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06e949a0-1f23-46bf-b910-fb634889464f (operator) has been started and output is visible here.\u001b[0m\n\nPLAY [Make ssh pipelining working] *********************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 13 January 2026  00:22:37 +0000 (0:00:00.135)       0:00:00.135 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Do not require tty for all users] ****************************************\nTuesday 13 January 2026  00:22:41 +0000 (0:00:04.319)       0:00:04.455 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role operator] *****************************************************\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nTuesday 13 January 2026  00:22:42 +0000 (0:00:00.766)       0:00:05.221 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nTuesday 13 January 2026  00:22:42 +0000 (0:00:00.184)       0:00:05.405 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create operator group] **************************\nTuesday 13 January 2026  00:22:42 +0000 (0:00:00.170)       0:00:05.576 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.operator : Create user] ************************************\nTuesday 13 January 2026  00:22:43 +0000 (0:00:00.631)       0:00:06.208 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nTuesday 13 January 2026  00:22:44 +0000 (0:00:00.749)       0:00:06.957 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nTuesday 13 January 2026  00:22:45 +0000 (0:00:01.220)       0:00:08.177 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nTuesday 13 January 2026  00:22:46 +0000 (0:00:01.165)       0:00:09.343 ******* \n\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m\n\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m\n\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nTuesday 13 January 2026  00:22:47 +0000 (0:00:01.204)       0:00:10.547 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nTuesday 13 January 2026  00:22:47 +0000 (0:00:00.152)       0:00:10.699 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nTuesday 13 January 2026  00:22:48 +0000 (0:00:00.169)       0:00:10.869 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nTuesday 13 January 2026  00:22:48 +0000 (0:00:00.592)       0:00:11.461 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nTuesday 13 January 2026  00:22:48 +0000 (0:00:00.151)       0:00:11.613 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nTuesday 13 January 2026  00:22:49 +0000 (0:00:00.686)       0:00:12.300 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nTuesday 13 January 2026  00:22:49 +0000 (0:00:00.152)       0:00:12.452 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nTuesday 13 January 2026  00:22:49 +0000 (0:00:00.150)       0:00:12.603 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set password] ***********************************\nTuesday 13 January 2026  00:22:49 +0000 (0:00:00.152)       0:00:12.756 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nTuesday 13 January 2026  00:22:50 +0000 (0:00:00.663)       0:00:13.419 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:22:50 +0000 (0:00:00.262)       0:00:13.682 ******* \n=============================================================================== \nGathering Facts --------------------------------------------------------- 4.32s\nosism.commons.operator : Add user to additional groups ------------------ 1.22s\nosism.commons.operator : Set language variables in .bashrc configuration file --- 1.20s\nosism.commons.operator : Copy user sudoers file ------------------------- 1.17s\nDo not require tty for all users ---------------------------------------- 0.77s\nosism.commons.operator : Create user ------------------------------------ 0.75s\nosism.commons.operator : Set ssh authorized keys ------------------------ 0.69s\nosism.commons.operator : Set password ----------------------------------- 0.66s\nosism.commons.operator : Create operator group -------------------------- 0.63s\nosism.commons.operator : Create .ssh directory -------------------------- 0.59s\nosism.commons.operator : Unset & lock password -------------------------- 0.26s\nosism.commons.operator : Gather variables for each operating system ----- 0.18s\nosism.commons.operator : Set operator_groups variable to default value --- 0.17s\nosism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.17s\nosism.commons.operator : Delete authorized GitHub accounts -------------- 0.15s\nosism.commons.operator : Delete ssh authorized keys --------------------- 0.15s\nosism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.15s\nosism.commons.operator : Check number of SSH authorized keys ------------ 0.15s\nosism.commons.operator : Set authorized GitHub accounts ----------------- 0.15s\n+ osism apply --environment custom facts\n\u001b[32m2026-01-13 00:22:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m\n\u001b[32m2026-01-13 00:23:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 680fbef7-b83d-4dcb-b9b6-a0e0d6048bcb (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:23:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 680fbef7-b83d-4dcb-b9b6-a0e0d6048bcb (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy custom network devices fact] ****************************************\n\nTASK [Create custom facts directory] *******************************************\nTuesday 13 January 2026  00:23:07 +0000 (0:00:00.073)       0:00:00.073 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Copy fact file] **********************************************************\nTuesday 13 January 2026  00:23:08 +0000 (0:00:01.419)       0:00:01.493 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Copy custom ceph devices facts] ******************************************\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nTuesday 13 January 2026  00:23:09 +0000 (0:00:01.089)       0:00:02.583 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nTuesday 13 January 2026  00:23:10 +0000 (0:00:00.078)       0:00:02.661 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nTuesday 13 January 2026  00:23:10 +0000 (0:00:00.182)       0:00:02.844 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nTuesday 13 January 2026  00:23:10 +0000 (0:00:00.183)       0:00:03.027 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nTuesday 13 January 2026  00:23:10 +0000 (0:00:00.127)       0:00:03.155 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nTuesday 13 January 2026  00:23:10 +0000 (0:00:00.424)       0:00:03.579 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nTuesday 13 January 2026  00:23:11 +0000 (0:00:00.112)       0:00:03.692 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nTuesday 13 January 2026  00:23:12 +0000 (0:00:00.997)       0:00:04.689 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nTuesday 13 January 2026  00:23:12 +0000 (0:00:00.464)       0:00:05.154 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nTuesday 13 January 2026  00:23:13 +0000 (0:00:01.046)       0:00:06.200 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Install required packages (RedHat)] **************************************\nTuesday 13 January 2026  00:23:29 +0000 (0:00:15.507)       0:00:21.708 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (Debian)] **************************************\nTuesday 13 January 2026  00:23:29 +0000 (0:00:00.100)       0:00:21.808 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Create custom facts directory] *******************************************\nTuesday 13 January 2026  00:23:37 +0000 (0:00:08.343)       0:00:30.152 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Copy fact files] *********************************************************\nTuesday 13 January 2026  00:23:37 +0000 (0:00:00.458)       0:00:30.610 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nTuesday 13 January 2026  00:23:41 +0000 (0:00:03.568)       0:00:34.179 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:23:42 +0000 (0:00:01.266)       0:00:35.446 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:23:46 +0000 (0:00:03.759)       0:00:39.205 ******* \n=============================================================================== \nosism.commons.repository : Update package cache ------------------------ 15.51s\nInstall required packages (Debian) -------------------------------------- 8.34s\nGathers facts about hosts ----------------------------------------------- 3.76s\nCopy fact files --------------------------------------------------------- 3.57s\nCreate custom facts directory ------------------------------------------- 1.42s\nosism.commons.repository : Force update of package cache ---------------- 1.27s\nCopy fact file ---------------------------------------------------------- 1.09s\nosism.commons.repository : Copy ubuntu.sources file --------------------- 1.05s\nosism.commons.repository : Copy 99osism apt configuration --------------- 1.00s\nosism.commons.repository : Remove sources.list file --------------------- 0.46s\nCreate custom facts directory ------------------------------------------- 0.46s\nosism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.42s\nosism.commons.repository : Set repositories to default ------------------ 0.18s\nosism.commons.repository : Set repository_default fact to default value --- 0.18s\nosism.commons.repository : Include distribution specific repository tasks --- 0.13s\nosism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.11s\nInstall required packages (RedHat) -------------------------------------- 0.10s\nosism.commons.repository : Gather variables for each operating system --- 0.08s\n+ osism apply bootstrap\n\u001b[32m2026-01-13 00:23:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b55d8260-af86-4169-9c84-5a2c3e1c880b (bootstrap) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:23:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b55d8260-af86-4169-9c84-5a2c3e1c880b (bootstrap) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on state bootstrap] ************************************\n\nTASK [Group hosts based on state bootstrap] ************************************\nTuesday 13 January 2026  00:24:03 +0000 (0:00:00.114)       0:00:00.114 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:24:03 +0000 (0:00:00.197)       0:00:00.312 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts (if using --limit)] ***************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:24:08 +0000 (0:00:04.697)       0:00:05.009 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 1] ********************************************\n\nTASK [osism.commons.hostname : Set hostname] ***********************************\nTuesday 13 January 2026  00:24:08 +0000 (0:00:00.454)       0:00:05.464 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.hostname : Copy /etc/hostname] *****************************\nTuesday 13 January 2026  00:24:10 +0000 (0:00:01.345)       0:00:06.809 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.hosts : Include type specific tasks] ***********************\nTuesday 13 January 2026  00:24:11 +0000 (0:00:01.230)       0:00:08.039 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.hosts : Copy /etc/hosts file] ******************************\nTuesday 13 January 2026  00:24:11 +0000 (0:00:00.263)       0:00:08.303 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.proxy : Include distribution specific tasks] ***************\nTuesday 13 January 2026  00:24:13 +0000 (0:00:02.092)       0:00:10.395 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.proxy : Configure proxy parameters for apt] ****************\nTuesday 13 January 2026  00:24:13 +0000 (0:00:00.250)       0:00:10.646 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.proxy : Set system wide settings in environment file] ******\nTuesday 13 January 2026  00:24:15 +0000 (0:00:01.028)       0:00:11.674 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.proxy : Remove system wide settings in environment file] ***\nTuesday 13 January 2026  00:24:15 +0000 (0:00:00.575)       0:00:12.250 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nTuesday 13 January 2026  00:24:15 +0000 (0:00:00.424)       0:00:12.674 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nTuesday 13 January 2026  00:24:16 +0000 (0:00:00.225)       0:00:12.900 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nTuesday 13 January 2026  00:24:16 +0000 (0:00:00.281)       0:00:13.182 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nTuesday 13 January 2026  00:24:16 +0000 (0:00:00.297)       0:00:13.479 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nTuesday 13 January 2026  00:24:18 +0000 (0:00:01.637)       0:00:15.117 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nTuesday 13 January 2026  00:24:18 +0000 (0:00:00.245)       0:00:15.363 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nTuesday 13 January 2026  00:24:19 +0000 (0:00:00.552)       0:00:15.915 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nTuesday 13 January 2026  00:24:19 +0000 (0:00:00.334)       0:00:16.250 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nTuesday 13 January 2026  00:24:20 +0000 (0:00:00.539)       0:00:16.791 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nTuesday 13 January 2026  00:24:21 +0000 (0:00:01.146)       0:00:17.937 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nTuesday 13 January 2026  00:24:22 +0000 (0:00:01.170)       0:00:19.108 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nTuesday 13 January 2026  00:24:22 +0000 (0:00:00.319)       0:00:19.427 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nTuesday 13 January 2026  00:24:24 +0000 (0:00:01.272)       0:00:20.700 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nTuesday 13 January 2026  00:24:24 +0000 (0:00:00.228)       0:00:20.928 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nTuesday 13 January 2026  00:24:24 +0000 (0:00:00.233)       0:00:21.161 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nTuesday 13 January 2026  00:24:24 +0000 (0:00:00.228)       0:00:21.390 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nTuesday 13 January 2026  00:24:25 +0000 (0:00:00.290)       0:00:21.681 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nTuesday 13 January 2026  00:24:25 +0000 (0:00:00.503)       0:00:22.185 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nTuesday 13 January 2026  00:24:25 +0000 (0:00:00.253)       0:00:22.438 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nTuesday 13 January 2026  00:24:26 +0000 (0:00:00.965)       0:00:23.403 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nTuesday 13 January 2026  00:24:27 +0000 (0:00:00.523)       0:00:23.927 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nTuesday 13 January 2026  00:24:28 +0000 (0:00:01.157)       0:00:25.084 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Gather variables for each operating system] *****\nTuesday 13 January 2026  00:24:44 +0000 (0:00:16.143)       0:00:41.228 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****\nTuesday 13 January 2026  00:24:44 +0000 (0:00:00.240)       0:00:41.468 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***\nTuesday 13 January 2026  00:24:45 +0000 (0:00:00.232)       0:00:41.701 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include distribution specific install tasks] ****\nTuesday 13 January 2026  00:24:45 +0000 (0:00:00.217)       0:00:41.919 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Install rsyslog package] ************************\nTuesday 13 January 2026  00:24:45 +0000 (0:00:00.263)       0:00:42.182 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********\nTuesday 13 January 2026  00:24:47 +0000 (0:00:01.870)       0:00:44.053 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Manage rsyslog service] *************************\nTuesday 13 January 2026  00:24:48 +0000 (0:00:01.189)       0:00:45.243 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include fluentd tasks] **************************\nTuesday 13 January 2026  00:24:49 +0000 (0:00:00.869)       0:00:46.112 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***\nTuesday 13 January 2026  00:24:49 +0000 (0:00:00.286)       0:00:46.399 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.rsyslog : Include additional log server tasks] ************\nTuesday 13 January 2026  00:24:50 +0000 (0:00:01.106)       0:00:47.505 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include logrotate tasks] ************************\nTuesday 13 January 2026  00:24:51 +0000 (0:00:00.240)       0:00:47.746 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Ensure logrotate package is installed] **********\nTuesday 13 January 2026  00:24:51 +0000 (0:00:00.292)       0:00:48.039 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************\nTuesday 13 January 2026  00:24:53 +0000 (0:00:01.825)       0:00:49.865 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.systohc : Install util-linux-extra package] ****************\nTuesday 13 January 2026  00:24:54 +0000 (0:00:01.116)       0:00:50.981 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.systohc : Sync hardware clock] *****************************\nTuesday 13 January 2026  00:25:06 +0000 (0:00:11.698)       0:01:02.679 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.configfs : Start sys-kernel-config mount] ******************\nTuesday 13 January 2026  00:25:07 +0000 (0:00:01.656)       0:01:04.335 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Gather variables for each operating system] *****\nTuesday 13 January 2026  00:25:08 +0000 (0:00:00.887)       0:01:05.223 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***\nTuesday 13 January 2026  00:25:08 +0000 (0:00:00.210)       0:01:05.434 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Include distribution specific package tasks] ****\nTuesday 13 January 2026  00:25:08 +0000 (0:00:00.217)       0:01:05.651 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.packages : Install needrestart package] ********************\nTuesday 13 January 2026  00:25:09 +0000 (0:00:00.266)       0:01:05.917 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Set needrestart mode] ***************************\nTuesday 13 January 2026  00:25:10 +0000 (0:00:01.667)       0:01:07.585 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***\nTuesday 13 January 2026  00:25:11 +0000 (0:00:00.587)       0:01:08.173 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Update package cache] ***************************\nTuesday 13 January 2026  00:25:11 +0000 (0:00:00.234)       0:01:08.408 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Download upgrade packages] **********************\nTuesday 13 January 2026  00:25:13 +0000 (0:00:01.314)       0:01:09.722 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Upgrade packages] *******************************\nTuesday 13 January 2026  00:25:14 +0000 (0:00:01.735)       0:01:11.458 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Download required packages] *********************\nTuesday 13 January 2026  00:25:17 +0000 (0:00:02.752)       0:01:14.210 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.packages : Install required packages] **********************\nTuesday 13 January 2026  00:25:55 +0000 (0:00:38.022)       0:01:52.233 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Remove useless packages from the cache] *********\nTuesday 13 January 2026  00:27:12 +0000 (0:01:17.350)       0:03:09.583 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Remove dependencies that are no longer required] ***\nTuesday 13 January 2026  00:27:14 +0000 (0:00:01.891)       0:03:11.475 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sysctl : Include sysctl tasks] *****************************\nTuesday 13 January 2026  00:27:24 +0000 (0:00:10.154)       0:03:21.630 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********\nTuesday 13 January 2026  00:27:25 +0000 (0:00:00.362)       0:03:21.992 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************\nTuesday 13 January 2026  00:27:25 +0000 (0:00:00.608)       0:03:22.601 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************\nTuesday 13 January 2026  00:27:33 +0000 (0:00:07.949)       0:03:30.550 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************\nTuesday 13 January 2026  00:27:34 +0000 (0:00:00.637)       0:03:31.188 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on network] *****************\nTuesday 13 January 2026  00:27:35 +0000 (0:00:00.617)       0:03:31.806 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************\nTuesday 13 January 2026  00:27:36 +0000 (0:00:01.604)       0:03:33.410 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\nTASK [osism.commons.limits : Include limits tasks] *****************************\nTuesday 13 January 2026  00:27:37 +0000 (0:00:00.602)       0:03:34.012 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Populate service facts] *************************\nTuesday 13 January 2026  00:27:37 +0000 (0:00:00.327)       0:03:34.340 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.services : Check services] *********************************\nTuesday 13 January 2026  00:27:43 +0000 (0:00:05.539)       0:03:39.879 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Start/enable required services] *****************\nTuesday 13 January 2026  00:27:43 +0000 (0:00:00.328)       0:03:40.208 ******* \n\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m\n\nTASK [osism.commons.motd : Include distribution specific configure tasks] ******\nTuesday 13 January 2026  00:27:44 +0000 (0:00:01.292)       0:03:41.501 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.motd : Remove update-motd package] *************************\nTuesday 13 January 2026  00:27:45 +0000 (0:00:00.517)       0:03:42.018 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************\nTuesday 13 January 2026  00:27:46 +0000 (0:00:01.265)       0:03:43.284 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Disable the dynamic motd-news service] **************\nTuesday 13 January 2026  00:27:47 +0000 (0:00:00.673)       0:03:43.957 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********\nTuesday 13 January 2026  00:27:47 +0000 (0:00:00.636)       0:03:44.594 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Remove pam_motd.so rule] ****************************\nTuesday 13 January 2026  00:27:48 +0000 (0:00:00.666)       0:03:45.261 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262699.1741004, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262720.6874278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262712.4412637, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262727.7672517, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262721.3498943, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262703.8333607, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262705.3500478, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [osism.commons.motd : Copy motd file] *************************************\nTuesday 13 January 2026  00:27:49 +0000 (0:00:01.020)       0:03:46.281 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue file] ************************************\nTuesday 13 January 2026  00:27:50 +0000 (0:00:01.256)       0:03:47.538 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue.net file] ********************************\nTuesday 13 January 2026  00:27:52 +0000 (0:00:01.155)       0:03:48.693 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to print the motd] ********************\nTuesday 13 January 2026  00:27:53 +0000 (0:00:01.244)       0:03:49.938 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to not print the motd] ****************\nTuesday 13 January 2026  00:27:53 +0000 (0:00:00.290)       0:03:50.228 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rng : Include distribution specific install tasks] ********\nTuesday 13 January 2026  00:27:54 +0000 (0:00:00.728)       0:03:50.957 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rng : Install rng package] ********************************\nTuesday 13 January 2026  00:27:54 +0000 (0:00:00.419)       0:03:51.377 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.rng : Remove haveged package] *****************************\nTuesday 13 January 2026  00:28:03 +0000 (0:00:08.546)       0:03:59.923 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rng : Manage rng service] *********************************\nTuesday 13 January 2026  00:28:04 +0000 (0:00:01.511)       0:04:01.434 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Gather variables for each operating system] ******\nTuesday 13 January 2026  00:28:05 +0000 (0:00:01.231)       0:04:02.665 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***\nTuesday 13 January 2026  00:28:06 +0000 (0:00:00.324)       0:04:02.990 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***\nTuesday 13 January 2026  00:28:06 +0000 (0:00:00.347)       0:04:03.338 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Populate service facts] **************************\nTuesday 13 January 2026  00:28:06 +0000 (0:00:00.318)       0:04:03.657 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Include distribution specific timer tasks] *******\nTuesday 13 January 2026  00:28:11 +0000 (0:00:04.886)       0:04:08.543 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Disable apt-daily timers] ************************\nTuesday 13 January 2026  00:28:12 +0000 (0:00:00.464)       0:04:09.008 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include service tasks] ***************************\nTuesday 13 January 2026  00:28:12 +0000 (0:00:00.357)       0:04:09.365 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup services] ********************************\nTuesday 13 January 2026  00:28:13 +0000 (0:00:00.407)       0:04:09.773 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include packages tasks] **************************\nTuesday 13 January 2026  00:28:13 +0000 (0:00:00.331)       0:04:10.105 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup installed packages] **********************\nTuesday 13 January 2026  00:28:13 +0000 (0:00:00.401)       0:04:10.507 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloudinit package] ************************\nTuesday 13 January 2026  00:28:48 +0000 (0:00:34.302)       0:04:44.809 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********\nTuesday 13 January 2026  00:28:56 +0000 (0:00:08.842)       0:04:53.652 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove useless packages from the cache] **********\nTuesday 13 January 2026  00:29:05 +0000 (0:00:08.171)       0:05:01.823 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***\nTuesday 13 January 2026  00:29:07 +0000 (0:00:01.923)       0:05:03.747 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.cleanup : Include cloudinit tasks] *************************\nTuesday 13 January 2026  00:29:13 +0000 (0:00:06.615)       0:05:10.362 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******\nTuesday 13 January 2026  00:29:14 +0000 (0:00:00.620)       0:05:10.983 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Install tzdata package] *************************\nTuesday 13 January 2026  00:29:15 +0000 (0:00:00.822)       0:05:11.805 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.timezone : Set timezone to UTC] ****************************\nTuesday 13 January 2026  00:29:16 +0000 (0:00:01.768)       0:05:13.574 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.timezone : Create /etc/adjtime file] ***********************\nTuesday 13 January 2026  00:29:17 +0000 (0:00:00.810)       0:05:14.385 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************\nTuesday 13 January 2026  00:29:18 +0000 (0:00:00.295)       0:05:14.680 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Gather variables for each operating system] ******\nTuesday 13 January 2026  00:29:18 +0000 (0:00:00.401)       0:05:15.081 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_version variable to default value] ****\nTuesday 13 January 2026  00:29:18 +0000 (0:00:00.280)       0:05:15.362 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_cli_version variable to default value] ***\nTuesday 13 January 2026  00:29:18 +0000 (0:00:00.269)       0:05:15.631 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Print used docker version] ***********************\nTuesday 13 January 2026  00:29:19 +0000 (0:00:00.280)       0:05:15.912 ******* \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Print used docker cli version] *******************\nTuesday 13 January 2026  00:29:19 +0000 (0:00:00.290)       0:05:16.203 ******* \n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Include block storage tasks] *********************\nTuesday 13 January 2026  00:29:19 +0000 (0:00:00.292)       0:05:16.495 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include zram storage tasks] **********************\nTuesday 13 January 2026  00:29:20 +0000 (0:00:00.267)       0:05:16.763 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include docker install tasks] ********************\nTuesday 13 January 2026  00:29:20 +0000 (0:00:00.306)       0:05:17.070 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Remove old architecture-dependent repository] ****\nTuesday 13 January 2026  00:29:20 +0000 (0:00:00.409)       0:05:17.479 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Gather package facts] ****************************\nTuesday 13 January 2026  00:29:21 +0000 (0:00:00.982)       0:05:18.462 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.docker : Check whether packages are installed that should not be installed] ***\nTuesday 13 January 2026  00:29:24 +0000 (0:00:02.877)       0:05:21.339 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install apt-transport-https package] *************\nTuesday 13 January 2026  00:29:25 +0000 (0:00:00.566)       0:05:21.906 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Add repository gpg key] **************************\nTuesday 13 January 2026  00:29:32 +0000 (0:00:07.060)       0:05:28.967 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Add repository] **********************************\nTuesday 13 January 2026  00:29:33 +0000 (0:00:01.162)       0:05:30.129 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Update package cache] ****************************\nTuesday 13 January 2026  00:29:41 +0000 (0:00:08.540)       0:05:38.669 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Pin docker package version] **********************\nTuesday 13 January 2026  00:29:45 +0000 (0:00:03.441)       0:05:42.111 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Pin docker-cli package version] ******************\nTuesday 13 January 2026  00:29:46 +0000 (0:00:01.401)       0:05:43.512 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Unlock containerd package] ***********************\nTuesday 13 January 2026  00:29:48 +0000 (0:00:01.520)       0:05:45.033 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install containerd package] **********************\nTuesday 13 January 2026  00:29:48 +0000 (0:00:00.627)       0:05:45.660 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.docker : Lock containerd package] *************************\nTuesday 13 January 2026  00:29:59 +0000 (0:00:10.733)       0:05:56.394 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker-cli package] **********************\nTuesday 13 January 2026  00:30:00 +0000 (0:00:00.904)       0:05:57.299 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker package] **************************\nTuesday 13 January 2026  00:30:11 +0000 (0:00:10.399)       0:06:07.699 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Unblock installation of python docker packages] ***\nTuesday 13 January 2026  00:30:23 +0000 (0:00:12.549)       0:06:20.248 ******* \n\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package] ******************\nTuesday 13 January 2026  00:30:24 +0000 (0:00:01.334)       0:06:21.583 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package from Debian Sid] ***\nTuesday 13 January 2026  00:30:25 +0000 (0:00:00.521)       0:06:22.104 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***\nTuesday 13 January 2026  00:30:29 +0000 (0:00:04.013)       0:06:26.118 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***\nTuesday 13 January 2026  00:30:29 +0000 (0:00:00.494)       0:06:26.612 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***\nTuesday 13 January 2026  00:30:30 +0000 (0:00:00.714)       0:06:27.327 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***\nTuesday 13 January 2026  00:30:31 +0000 (0:00:00.489)       0:06:27.816 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install packages required by docker login] *******\nTuesday 13 January 2026  00:30:31 +0000 (0:00:00.478)       0:06:28.295 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Ensure that some packages are not installed] *****\nTuesday 13 January 2026  00:30:32 +0000 (0:00:00.531)       0:06:28.827 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include config tasks] ****************************\nTuesday 13 January 2026  00:30:34 +0000 (0:00:02.211)       0:06:31.038 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create plugins directory] ************************\nTuesday 13 January 2026  00:30:35 +0000 (0:00:00.825)       0:06:31.864 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Create systemd overlay directory] ****************\nTuesday 13 January 2026  00:30:36 +0000 (0:00:00.848)       0:06:32.712 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Copy systemd overlay file] ***********************\nTuesday 13 January 2026  00:30:36 +0000 (0:00:00.868)       0:06:33.581 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***\nTuesday 13 January 2026  00:30:38 +0000 (0:00:01.517)       0:06:35.099 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy limits configuration file] ******************\nTuesday 13 January 2026  00:30:39 +0000 (0:00:01.480)       0:06:36.579 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy daemon.json configuration file] *************\nTuesday 13 January 2026  00:30:41 +0000 (0:00:01.282)       0:06:37.862 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include service tasks] ***************************\nTuesday 13 January 2026  00:30:42 +0000 (0:00:01.413)       0:06:39.276 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon] ***************************\nTuesday 13 January 2026  00:30:43 +0000 (0:00:01.003)       0:06:40.279 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage service] **********************************\nTuesday 13 January 2026  00:30:45 +0000 (0:00:01.420)       0:06:41.699 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage docker socket service] ********************\nTuesday 13 January 2026  00:30:46 +0000 (0:00:01.211)       0:06:42.911 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Manage containerd service] ***********************\nTuesday 13 January 2026  00:30:47 +0000 (0:00:01.182)       0:06:44.093 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include bootstrap tasks] *************************\nTuesday 13 January 2026  00:30:48 +0000 (0:00:01.309)       0:06:45.403 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.860)       0:06:46.264 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.039)       0:06:46.304 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.037)       0:06:46.342 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.044)       0:06:46.386 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.038)       0:06:46.424 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.037)       0:06:46.462 ******* \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.044)       0:06:46.507 ******* \n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nTuesday 13 January 2026  00:30:49 +0000 (0:00:00.038)       0:06:46.545 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************\nTuesday 13 January 2026  00:30:51 +0000 (0:00:01.380)       0:06:47.926 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********\nTuesday 13 January 2026  00:30:52 +0000 (0:00:01.482)       0:06:49.408 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Restart docker service] ***************\nTuesday 13 January 2026  00:30:54 +0000 (0:00:01.423)       0:06:50.832 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****\nTuesday 13 January 2026  00:30:56 +0000 (0:00:02.309)       0:06:53.141 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Add user to docker group] ************************\nTuesday 13 January 2026  00:30:56 +0000 (0:00:00.105)       0:06:53.247 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Log into private registry and force re-authorization] ***\nTuesday 13 January 2026  00:30:57 +0000 (0:00:01.066)       0:06:54.313 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include facts tasks] *****************************\nTuesday 13 January 2026  00:30:58 +0000 (0:00:00.506)       0:06:54.819 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create facts directory] **************************\nTuesday 13 January 2026  00:30:59 +0000 (0:00:01.037)       0:06:55.857 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy docker fact files] **************************\nTuesday 13 January 2026  00:31:00 +0000 (0:00:00.837)       0:06:56.695 ******* \n\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m\n\nTASK [osism.commons.docker_compose : This install type is not supported] *******\nTuesday 13 January 2026  00:31:02 +0000 (0:00:02.566)       0:06:59.261 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Include distribution specific install tasks] ***\nTuesday 13 January 2026  00:31:03 +0000 (0:00:00.634)       0:06:59.895 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***\nTuesday 13 January 2026  00:31:03 +0000 (0:00:00.771)       0:07:00.667 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******\nTuesday 13 January 2026  00:31:04 +0000 (0:00:00.828)       0:07:01.495 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose binary] *************\nTuesday 13 January 2026  00:31:05 +0000 (0:00:01.033)       0:07:02.528 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Uninstall docker-compose package] *********\nTuesday 13 January 2026  00:31:06 +0000 (0:00:00.472)       0:07:03.001 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose script] ***************\nTuesday 13 January 2026  00:31:07 +0000 (0:00:01.625)       0:07:04.626 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****\nTuesday 13 January 2026  00:31:08 +0000 (0:00:00.481)       0:07:05.108 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********\nTuesday 13 January 2026  00:31:16 +0000 (0:00:08.415)       0:07:13.523 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Enable osism.target] **********************\nTuesday 13 January 2026  00:31:18 +0000 (0:00:01.527)       0:07:15.051 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****\nTuesday 13 January 2026  00:31:20 +0000 (0:00:01.890)       0:07:16.942 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 13 January 2026  00:31:21 +0000 (0:00:01.637)       0:07:18.579 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 13 January 2026  00:31:22 +0000 (0:00:00.952)       0:07:19.531 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check minimum and maximum number of servers] *****\nTuesday 13 January 2026  00:31:23 +0000 (0:00:00.969)       0:07:20.501 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Gather variables for each operating system] ******\nTuesday 13 January 2026  00:31:24 +0000 (0:00:00.552)       0:07:21.053 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***\nTuesday 13 January 2026  00:31:24 +0000 (0:00:00.512)       0:07:21.566 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_key_file variable to default value] ***\nTuesday 13 January 2026  00:31:25 +0000 (0:00:00.530)       0:07:22.096 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Populate service facts] **************************\nTuesday 13 January 2026  00:31:26 +0000 (0:00:00.712)       0:07:22.808 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Manage timesyncd service] ************************\nTuesday 13 January 2026  00:31:31 +0000 (0:00:05.550)       0:07:28.359 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Include distribution specific install tasks] *****\nTuesday 13 January 2026  00:31:32 +0000 (0:00:00.608)       0:07:28.967 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.chrony : Install package] *********************************\nTuesday 13 January 2026  00:31:33 +0000 (0:00:01.134)       0:07:30.102 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Manage chrony service] ***************************\nTuesday 13 January 2026  00:31:35 +0000 (0:00:01.982)       0:07:32.084 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check if configuration file exists] **************\nTuesday 13 January 2026  00:31:36 +0000 (0:00:01.186)       0:07:33.271 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Copy configuration file] *************************\nTuesday 13 January 2026  00:31:37 +0000 (0:00:00.854)       0:07:34.126 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\nTASK [osism.services.lldpd : Include distribution specific install tasks] ******\nTuesday 13 January 2026  00:31:39 +0000 (0:00:01.937)       0:07:36.063 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.lldpd : Install lldpd package] ****************************\nTuesday 13 January 2026  00:31:40 +0000 (0:00:00.900)       0:07:36.963 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.lldpd : Manage lldpd service] *****************************\nTuesday 13 January 2026  00:31:49 +0000 (0:00:09.644)       0:07:46.607 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********\nTuesday 13 January 2026  00:31:51 +0000 (0:00:01.880)       0:07:48.488 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************\nTuesday 13 January 2026  00:31:53 +0000 (0:00:01.367)       0:07:49.855 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap role part 2] *********************************************\n\nTASK [Include hardening role] **************************************************\nTuesday 13 January 2026  00:31:54 +0000 (0:00:01.277)       0:07:51.133 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 3] ********************************************\n\nTASK [osism.services.journald : Copy configuration file] ***********************\nTuesday 13 January 2026  00:31:55 +0000 (0:00:00.668)       0:07:51.802 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.journald : Manage journald service] ***********************\nTuesday 13 January 2026  00:31:56 +0000 (0:00:01.433)       0:07:53.235 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include auditd role] *****************************************************\nTuesday 13 January 2026  00:31:57 +0000 (0:00:01.414)       0:07:54.650 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include smartd role] *****************************************************\nTuesday 13 January 2026  00:31:58 +0000 (0:00:00.475)       0:07:55.125 ******* \n\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Include distribution specific install tasks] *****\nTuesday 13 January 2026  00:31:59 +0000 (0:00:00.918)       0:07:56.043 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Install smartmontools package] *******************\nTuesday 13 January 2026  00:32:00 +0000 (0:00:00.758)       0:07:56.802 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.smartd : Create /var/log/smartd directory] ****************\nTuesday 13 January 2026  00:32:09 +0000 (0:00:09.030)       0:08:05.832 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.smartd : Copy smartmontools configuration file] ***********\nTuesday 13 January 2026  00:32:10 +0000 (0:00:01.053)       0:08:06.886 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Manage smartd service] ***************************\nTuesday 13 January 2026  00:32:11 +0000 (0:00:01.344)       0:08:08.231 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.journald : Restart journald service] ***********\nTuesday 13 January 2026  00:32:13 +0000 (0:00:01.861)       0:08:10.093 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************\nTuesday 13 January 2026  00:32:14 +0000 (0:00:01.425)       0:08:11.518 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Set state bootstrap] *****************************************************\n\nTASK [Set osism.bootstrap.status fact] *****************************************\nTuesday 13 January 2026  00:32:15 +0000 (0:00:01.103)       0:08:12.621 ******* \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nTuesday 13 January 2026  00:32:16 +0000 (0:00:00.782)       0:08:13.404 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nTuesday 13 January 2026  00:32:17 +0000 (0:00:01.089)       0:08:14.494 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Set osism.bootstrap.timestamp fact] **************************************\nTuesday 13 January 2026  00:32:18 +0000 (0:00:01.122)       0:08:15.616 ******* \n\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nTuesday 13 January 2026  00:32:19 +0000 (0:00:00.795)       0:08:16.412 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nTuesday 13 January 2026  00:32:20 +0000 (0:00:01.081)       0:08:17.494 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:32:21 +0000 (0:00:01.072)       0:08:18.566 ******* \n=============================================================================== \nosism.commons.packages : Install required packages --------------------- 77.35s\nosism.commons.packages : Download required packages -------------------- 38.02s\nosism.commons.cleanup : Cleanup installed packages --------------------- 34.30s\nosism.commons.repository : Update package cache ------------------------ 16.14s\nosism.services.docker : Install docker package ------------------------- 12.55s\nosism.commons.systohc : Install util-linux-extra package --------------- 11.70s\nosism.services.docker : Install containerd package --------------------- 10.73s\nosism.services.docker : Install docker-cli package --------------------- 10.40s\nosism.commons.packages : Remove dependencies that are no longer required -- 10.15s\nosism.services.lldpd : Install lldpd package ---------------------------- 9.64s\nosism.services.smartd : Install smartmontools package ------------------- 9.03s\nosism.commons.cleanup : Remove cloudinit package ------------------------ 8.84s\nosism.services.rng : Install rng package -------------------------------- 8.55s\nosism.services.docker : Add repository ---------------------------------- 8.54s\nosism.commons.docker_compose : Install docker-compose-plugin package ---- 8.42s\nosism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.17s\nosism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 7.95s\nosism.services.docker : Install apt-transport-https package ------------- 7.06s\nosism.commons.cleanup : Remove dependencies that are no longer required --- 6.62s\nosism.services.chrony : Populate service facts -------------------------- 5.55s\n+ osism apply fail2ban\n\u001b[32m2026-01-13 00:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e47c4f7-4ec2-4007-886b-032cbee6e968 (fail2ban) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9e47c4f7-4ec2-4007-886b-032cbee6e968 (fail2ban) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role fail2ban] *****************************************************\n\nTASK [osism.services.fail2ban : Include distribution specific install tasks] ***\nTuesday 13 January 2026  00:32:39 +0000 (0:00:00.275)       0:00:00.275 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.fail2ban : Install fail2ban package] **********************\nTuesday 13 January 2026  00:32:40 +0000 (0:00:01.136)       0:00:01.411 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Copy configuration files] **********************\nTuesday 13 January 2026  00:32:52 +0000 (0:00:11.264)       0:00:12.676 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Manage fail2ban service] ***********************\nTuesday 13 January 2026  00:32:53 +0000 (0:00:01.506)       0:00:14.183 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Reload fail2ban configuration] *****************\nTuesday 13 January 2026  00:32:55 +0000 (0:00:01.581)       0:00:15.764 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:32:56 +0000 (0:00:01.654)       0:00:17.419 ******* \n=============================================================================== \nosism.services.fail2ban : Install fail2ban package --------------------- 11.26s\nosism.services.fail2ban : Reload fail2ban configuration ----------------- 1.65s\nosism.services.fail2ban : Manage fail2ban service ----------------------- 1.58s\nosism.services.fail2ban : Copy configuration files ---------------------- 1.51s\nosism.services.fail2ban : Include distribution specific install tasks --- 1.14s\n+ [[ -e /etc/redhat-release ]]\n+ osism apply network\n\u001b[32m2026-01-13 00:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08e96ea9-e9e0-489d-99af-6eaeeff5e332 (network) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08e96ea9-e9e0-489d-99af-6eaeeff5e332 (network) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role network] ******************************************************\n\nTASK [osism.commons.network : Gather variables for each operating system] ******\nTuesday 13 January 2026  00:33:13 +0000 (0:00:00.255)       0:00:00.255 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include type specific tasks] *********************\nTuesday 13 January 2026  00:33:14 +0000 (0:00:00.568)       0:00:00.823 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Install required packages] ***********************\nTuesday 13 January 2026  00:33:15 +0000 (0:00:01.005)       0:00:01.828 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove ifupdown package] *************************\nTuesday 13 January 2026  00:33:17 +0000 (0:00:02.040)       0:00:03.869 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Create required directories] *********************\nTuesday 13 January 2026  00:33:19 +0000 (0:00:01.735)       0:00:05.605 ******* \n\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m\n\nTASK [osism.commons.network : Prepare netplan configuration template] **********\nTuesday 13 January 2026  00:33:20 +0000 (0:00:00.957)       0:00:06.562 ******* \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Copy netplan configuration] **********************\nTuesday 13 January 2026  00:33:23 +0000 (0:00:03.244)       0:00:09.806 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove netplan configuration template] ***********\nTuesday 13 January 2026  00:33:24 +0000 (0:00:01.552)       0:00:11.359 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Check if path for interface file exists] *********\nTuesday 13 January 2026  00:33:26 +0000 (0:00:01.614)       0:00:12.974 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy interfaces file] ****************************\nTuesday 13 January 2026  00:33:27 +0000 (0:00:01.114)       0:00:14.088 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Install package networkd-dispatcher] *************\nTuesday 13 January 2026  00:33:28 +0000 (0:00:00.648)       0:00:14.737 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.network : Copy dispatcher scripts] *************************\nTuesday 13 January 2026  00:33:30 +0000 (0:00:02.203)       0:00:16.940 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m\n\nTASK [osism.commons.network : Manage service networkd-dispatcher] **************\nTuesday 13 January 2026  00:33:31 +0000 (0:00:00.902)       0:00:17.842 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include cleanup tasks] ***************************\nTuesday 13 January 2026  00:33:33 +0000 (0:00:01.627)       0:00:19.470 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nTuesday 13 January 2026  00:33:34 +0000 (0:00:01.193)       0:00:20.664 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Set network_configured_files fact] ***************\nTuesday 13 January 2026  00:33:35 +0000 (0:00:01.104)       0:00:21.769 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nTuesday 13 January 2026  00:33:36 +0000 (0:00:00.626)       0:00:22.395 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\nTASK [osism.commons.network : Include dummy interfaces] ************************\nTuesday 13 January 2026  00:33:37 +0000 (0:00:01.192)       0:00:23.588 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include vxlan interfaces] ************************\nTuesday 13 January 2026  00:33:37 +0000 (0:00:00.586)       0:00:24.175 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-0, testbed-node-2, testbed-manager, testbed-node-3, testbed-node-1, testbed-node-5, testbed-node-4\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd netdev files] ************\nTuesday 13 January 2026  00:33:42 +0000 (0:00:04.307)       0:00:28.482 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd network files] ***********\nTuesday 13 January 2026  00:33:47 +0000 (0:00:05.341)       0:00:33.824 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Include networkd cleanup tasks] ******************\nTuesday 13 January 2026  00:33:52 +0000 (0:00:05.033)       0:00:38.858 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nTuesday 13 January 2026  00:33:53 +0000 (0:00:01.071)       0:00:39.930 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nTuesday 13 January 2026  00:33:54 +0000 (0:00:01.010)       0:00:40.941 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include network extra init] **********************\nTuesday 13 January 2026  00:33:55 +0000 (0:00:00.783)       0:00:41.724 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init script] ****************\nTuesday 13 January 2026  00:33:56 +0000 (0:00:01.204)       0:00:42.928 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init systemd service] *******\nTuesday 13 January 2026  00:33:57 +0000 (0:00:00.632)       0:00:43.561 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Enable and start network-extra-init service] *****\nTuesday 13 January 2026  00:33:57 +0000 (0:00:00.740)       0:00:44.302 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Disable and stop network-extra-init service] *****\nTuesday 13 January 2026  00:33:58 +0000 (0:00:00.585)       0:00:44.887 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init systemd service] *******\nTuesday 13 January 2026  00:34:00 +0000 (0:00:01.753)       0:00:46.641 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init script] ****************\nTuesday 13 January 2026  00:34:01 +0000 (0:00:00.964)       0:00:47.606 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************\nTuesday 13 January 2026  00:34:03 +0000 (0:00:02.221)       0:00:49.827 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********\nTuesday 13 January 2026  00:34:04 +0000 (0:00:00.808)       0:00:50.635 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:34:04 +0000 (0:00:00.532)       0:00:51.168 ******* \n=============================================================================== \nosism.commons.network : Create systemd networkd netdev files ------------ 5.34s\nosism.commons.network : Create systemd networkd network files ----------- 5.03s\nosism.commons.network : Include vxlan interfaces ------------------------ 4.31s\nosism.commons.network : Prepare netplan configuration template ---------- 3.24s\nosism.commons.network : Remove network-extra-init script ---------------- 2.22s\nosism.commons.network : Install package networkd-dispatcher ------------- 2.20s\nosism.commons.network : Install required packages ----------------------- 2.04s\nosism.commons.network : Disable and stop network-extra-init service ----- 1.75s\nosism.commons.network : Remove ifupdown package ------------------------- 1.74s\nosism.commons.network : Manage service networkd-dispatcher -------------- 1.63s\nosism.commons.network : Remove netplan configuration template ----------- 1.61s\nosism.commons.network : Copy netplan configuration ---------------------- 1.55s\nosism.commons.network : Include network extra init ---------------------- 1.20s\nosism.commons.network : Include cleanup tasks --------------------------- 1.19s\nosism.commons.network : Remove unused configuration files --------------- 1.19s\nosism.commons.network : Check if path for interface file exists --------- 1.11s\nosism.commons.network : List existing configuration files --------------- 1.10s\nosism.commons.network : Include networkd cleanup tasks ------------------ 1.07s\nosism.commons.network : List existing configuration files --------------- 1.01s\nosism.commons.network : Include type specific tasks --------------------- 1.01s\n+ osism apply wireguard\n\u001b[32m2026-01-13 00:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8a352e9e-84ed-40db-bcd1-dca8ac701d87 (wireguard) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8a352e9e-84ed-40db-bcd1-dca8ac701d87 (wireguard) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role wireguard] ****************************************************\n\nTASK [osism.services.wireguard : Install iptables package] *********************\nTuesday 13 January 2026  00:34:21 +0000 (0:00:00.163)       0:00:00.163 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Install wireguard package] ********************\nTuesday 13 January 2026  00:34:22 +0000 (0:00:01.303)       0:00:01.467 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create public and private key - server] *******\nTuesday 13 January 2026  00:34:29 +0000 (0:00:06.612)       0:00:08.080 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create preshared key] *************************\nTuesday 13 January 2026  00:34:29 +0000 (0:00:00.561)       0:00:08.641 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get preshared key] ****************************\nTuesday 13 January 2026  00:34:30 +0000 (0:00:00.428)       0:00:09.070 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get public key - server] **********************\nTuesday 13 January 2026  00:34:31 +0000 (0:00:00.684)       0:00:09.754 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get private key - server] *********************\nTuesday 13 January 2026  00:34:31 +0000 (0:00:00.397)       0:00:10.151 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy wg0.conf configuration file] *************\nTuesday 13 January 2026  00:34:31 +0000 (0:00:00.405)       0:00:10.557 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy client configuration files] **************\nTuesday 13 January 2026  00:34:33 +0000 (0:00:01.162)       0:00:11.719 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********\nTuesday 13 January 2026  00:34:33 +0000 (0:00:00.954)       0:00:12.674 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************\nTuesday 13 January 2026  00:34:35 +0000 (0:00:01.621)       0:00:14.296 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:34:36 +0000 (0:00:00.899)       0:00:15.195 ******* \n=============================================================================== \nosism.services.wireguard : Install wireguard package -------------------- 6.61s\nosism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.62s\nosism.services.wireguard : Install iptables package --------------------- 1.30s\nosism.services.wireguard : Copy wg0.conf configuration file ------------- 1.16s\nosism.services.wireguard : Copy client configuration files -------------- 0.95s\nosism.services.wireguard : Restart wg0 service -------------------------- 0.90s\nosism.services.wireguard : Get preshared key ---------------------------- 0.68s\nosism.services.wireguard : Create public and private key - server ------- 0.56s\nosism.services.wireguard : Create preshared key ------------------------- 0.43s\nosism.services.wireguard : Get private key - server --------------------- 0.41s\nosism.services.wireguard : Get public key - server ---------------------- 0.40s\n+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0\r100    15  100    15    0     0    197      0 --:--:-- --:--:-- --:--:--   200\n+ osism apply --environment custom workarounds\n\u001b[32m2026-01-13 00:34:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m\n\u001b[32m2026-01-13 00:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8323183f-4a7a-455e-bd47-9ad302c4b40a (workarounds) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8323183f-4a7a-455e-bd47-9ad302c4b40a (workarounds) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on virtualization_role] ********************************\nTuesday 13 January 2026  00:34:53 +0000 (0:00:00.125)       0:00:00.125 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m\n\nPLAY [Apply netplan configuration on the manager node] *************************\n\nTASK [Apply netplan configuration] *********************************************\nTuesday 13 January 2026  00:34:54 +0000 (0:00:00.774)       0:00:00.899 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply netplan configuration on all other nodes] **************************\n\nTASK [Apply netplan configuration] *********************************************\nTuesday 13 January 2026  00:34:56 +0000 (0:00:02.139)       0:00:03.039 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Add custom CA certificates to non-manager nodes] *************************\n\nTASK [Copy custom CA certificates] *********************************************\nTuesday 13 January 2026  00:34:58 +0000 (0:00:01.996)       0:00:05.036 ******* \n\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\nTASK [Run update-ca-certificates] **********************************************\nTuesday 13 January 2026  00:34:59 +0000 (0:00:01.513)       0:00:06.549 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Run update-ca-trust] *****************************************************\nTuesday 13 January 2026  00:35:03 +0000 (0:00:04.052)       0:00:10.602 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Add a workaround service] ************************************************\n\nTASK [Copy workarounds.sh scripts] *********************************************\nTuesday 13 January 2026  00:35:04 +0000 (0:00:00.632)       0:00:11.234 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Copy workarounds systemd unit file] **************************************\nTuesday 13 January 2026  00:35:05 +0000 (0:00:01.459)       0:00:12.694 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Reload systemd daemon] ***************************************************\nTuesday 13 January 2026  00:35:07 +0000 (0:00:01.475)       0:00:14.169 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable workarounds.service (Debian)] *************************************\nTuesday 13 January 2026  00:35:08 +0000 (0:00:01.490)       0:00:15.659 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable and start workarounds.service (RedHat)] ***************************\nTuesday 13 January 2026  00:35:10 +0000 (0:00:01.708)       0:00:17.368 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************\n\nTASK [Install python3-docker] **************************************************\nTuesday 13 January 2026  00:35:11 +0000 (0:00:00.613)       0:00:17.982 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:35:13 +0000 (0:00:02.735)       0:00:20.717 ******* \n=============================================================================== \nRun update-ca-certificates ---------------------------------------------- 4.05s\nInstall python3-docker -------------------------------------------------- 2.74s\nApply netplan configuration --------------------------------------------- 2.14s\nApply netplan configuration --------------------------------------------- 2.00s\nEnable workarounds.service (Debian) ------------------------------------- 1.71s\nCopy custom CA certificates --------------------------------------------- 1.51s\nReload systemd daemon --------------------------------------------------- 1.49s\nCopy workarounds systemd unit file -------------------------------------- 1.48s\nCopy workarounds.sh scripts --------------------------------------------- 1.46s\nGroup hosts based on virtualization_role -------------------------------- 0.77s\nRun update-ca-trust ----------------------------------------------------- 0.63s\nEnable and start workarounds.service (RedHat) --------------------------- 0.61s\n+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-01-13 00:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dd3467b7-2d05-49ce-b4ed-d7ed65f1d10f (reboot) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dd3467b7-2d05-49ce-b4ed-d7ed65f1d10f (reboot) has been started and output is visible here.\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:30 +0000 (0:00:00.150)       0:00:00.150 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:30 +0000 (0:00:00.076)       0:00:00.227 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:31 +0000 (0:00:00.918)       0:00:01.145 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:31 +0000 (0:00:00.093)       0:00:01.238 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:31 +0000 (0:00:00.084)       0:00:01.323 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:32 +0000 (0:00:00.660)       0:00:01.983 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:32 +0000 (0:00:00.096)       0:00:02.079 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:32 +0000 (0:00:00.142)       0:00:02.221 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:33 +0000 (0:00:00.645)       0:00:02.867 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:33 +0000 (0:00:00.090)       0:00:02.958 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:33 +0000 (0:00:00.078)       0:00:03.037 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:34 +0000 (0:00:00.653)       0:00:03.690 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:34 +0000 (0:00:00.109)       0:00:03.799 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:34 +0000 (0:00:00.086)       0:00:03.886 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:35 +0000 (0:00:00.736)       0:00:04.623 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 13 January 2026  00:35:35 +0000 (0:00:00.117)       0:00:04.740 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 13 January 2026  00:35:35 +0000 (0:00:00.091)       0:00:04.832 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 13 January 2026  00:35:36 +0000 (0:00:00.696)       0:00:05.528 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:35:36 +0000 (0:00:00.034)       0:00:05.563 ******* \n=============================================================================== \nReboot system - do not wait for the reboot to complete ------------------ 4.31s\nExit playbook, if user did not mean to reboot systems ------------------- 0.56s\nReboot system - wait for the reboot to complete ------------------------- 0.54s\n+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-01-13 00:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 71b89266-deb6-462c-a198-3aa3a9843757 (wait-for-connection) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 71b89266-deb6-462c-a198-3aa3a9843757 (wait-for-connection) has been started and output is visible here.\u001b[0m\n\nPLAY [Wait until remote systems are reachable] *********************************\n\nTASK [Wait until remote system is reachable] ***********************************\nTuesday 13 January 2026  00:35:52 +0000 (0:00:00.169)       0:00:00.169 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:36:04 +0000 (0:00:11.477)       0:00:11.646 ******* \n=============================================================================== \nWait until remote system is reachable ---------------------------------- 11.48s\n+ osism apply hddtemp\n\u001b[32m2026-01-13 00:36:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ca78f17-8193-46a9-8856-5f474ebd664b (hddtemp) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:36:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7ca78f17-8193-46a9-8856-5f474ebd664b (hddtemp) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role hddtemp] ******************************************************\n\nTASK [osism.services.hddtemp : Gather variables for each operating system] *****\nTuesday 13 January 2026  00:36:21 +0000 (0:00:00.252)       0:00:00.252 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific install tasks] ****\nTuesday 13 January 2026  00:36:22 +0000 (0:00:00.684)       0:00:00.936 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Remove hddtemp package] *************************\nTuesday 13 January 2026  00:36:23 +0000 (0:00:01.155)       0:00:02.092 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************\nTuesday 13 January 2026  00:36:25 +0000 (0:00:02.214)       0:00:04.307 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Check if drivetemp module is available] *********\nTuesday 13 January 2026  00:36:26 +0000 (0:00:01.057)       0:00:05.364 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************\nTuesday 13 January 2026  00:36:27 +0000 (0:00:01.087)       0:00:06.451 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Install lm-sensors] *****************************\nTuesday 13 January 2026  00:36:28 +0000 (0:00:00.758)       0:00:07.210 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific service tasks] ****\nTuesday 13 January 2026  00:36:41 +0000 (0:00:13.623)       0:00:20.834 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Manage lm-sensors service] **********************\nTuesday 13 January 2026  00:36:43 +0000 (0:00:01.165)       0:00:22.000 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:36:44 +0000 (0:00:01.737)       0:00:23.738 ******* \n=============================================================================== \nosism.services.hddtemp : Install lm-sensors ---------------------------- 13.62s\nosism.services.hddtemp : Remove hddtemp package ------------------------- 2.21s\nosism.services.hddtemp : Manage lm-sensors service ---------------------- 1.74s\nosism.services.hddtemp : Include distribution specific service tasks ---- 1.17s\nosism.services.hddtemp : Include distribution specific install tasks ---- 1.16s\nosism.services.hddtemp : Check if drivetemp module is available --------- 1.09s\nosism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.06s\nosism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.76s\nosism.services.hddtemp : Gather variables for each operating system ----- 0.68s\n++ semver latest 7.1.1\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sudo systemctl restart manager.service\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\nARA in ceph-ansible already disabled.\nARA in kolla-ansible already disabled.\nARA in osism-ansible already disabled.\nARA in osism-kubernetes already disabled.\n+ osism apply gather-facts\n\u001b[32m2026-01-13 00:38:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ce44a44-dd4f-4cb9-89b4-2d880292c235 (gather-facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:38:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 6ce44a44-dd4f-4cb9-89b4-2d880292c235 (gather-facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:38:20 +0000 (0:00:00.186)       0:00:00.186 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 13 January 2026  00:38:28 +0000 (0:00:08.414)       0:00:08.600 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:38:29 +0000 (0:00:00.419)       0:00:09.019 ******* \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 8.41s\nGather facts for all hosts ---------------------------------------------- 0.42s\n+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper\n+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack\n+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring\n+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack\n+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry\n+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images\n+ [[ false == \\t\\r\\u\\e ]]",
                            "stdout_lines": [
                                "",
                                "# DEPLOY MANAGER",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY MANAGER'",
                                "+ echo",
                                "+ cat /opt/manager-vars.sh",
                                "export NUMBER_OF_NODES=6",
                                "",
                                "export CEPH_VERSION=reef",
                                "export CONFIGURATION_VERSION=main",
                                "export MANAGER_VERSION=latest",
                                "export OPENSTACK_VERSION=2024.2",
                                "",
                                "export ARA=false",
                                "export DEPLOY_MODE=manager",
                                "export TEMPEST=true",
                                "export IS_ZUUL=true",
                                "",
                                "export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "export EXTERNAL_API=false",
                                "",
                                "export IMAGE_USER=ubuntu",
                                "export IMAGE_NODE_USER=ubuntu",
                                "",
                                "export CEPH_STACK=ceph-ansible",
                                "",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver",
                                "+ docker version",
                                "Client: Docker Engine - Community",
                                " Version:           27.5.1",
                                " API version:       1.47",
                                " Go version:        go1.22.11",
                                " Git commit:        9f9e405",
                                " Built:             Wed Jan 22 13:41:48 2025",
                                " OS/Arch:           linux/amd64",
                                " Context:           default",
                                "",
                                "Server: Docker Engine - Community",
                                " Engine:",
                                "  Version:          27.5.1",
                                "  API version:      1.47 (minimum version 1.24)",
                                "  Go version:       go1.22.11",
                                "  Git commit:       4c9b3b0",
                                "  Built:            Wed Jan 22 13:41:48 2025",
                                "  OS/Arch:          linux/amd64",
                                "  Experimental:     false",
                                " containerd:",
                                "  Version:          v2.2.1",
                                "  GitCommit:        dea7da592f5d1d2b7755e3a161be07f43fad8f75",
                                " runc:",
                                "  Version:          1.3.4",
                                "  GitCommit:        v1.3.4-0-gd6d73eb8",
                                " docker-init:",
                                "  Version:          0.19.0",
                                "  GitCommit:        de40ad0",
                                "+ sh -c /opt/configuration/scripts/deploy/000-manager.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-ceph-version.sh reef",
                                "+ set -e",
                                "+ VERSION=reef",
                                "++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n ceph_version: reef ]]",
                                "+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ /opt/configuration/scripts/set-openstack-version.sh 2024.2",
                                "+ set -e",
                                "+ VERSION=2024.2",
                                "++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n openstack_version: 2024.2 ]]",
                                "+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ ceph-ansible == \\r\\o\\o\\k ]]",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo 'enable_osism_kubernetes: true'",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 2024.2 2025.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ /opt/configuration/scripts/enable-resource-nodes.sh",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml",
                                "",
                                "PLAY [Copy custom facts] *******************************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY [Before the deployment of the manager] ************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get /opt/manager-vars.sh] ************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add ara_server_mariadb_volume_type parameter] ****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Add netbox_enable parameter] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Use insecure glance configuration] ***************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Check if /etc/OTC_region exist] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add nova_compute_virt_type parameter] ************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Apply role traefik] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply traefik role] ******************************************************",
                                "included: osism.services.traefik for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/traefik)",
                                "changed: [testbed-manager] => (item=/opt/traefik/certificates)",
                                "changed: [testbed-manager] => (item=/opt/traefik/configuration)",
                                "",
                                "TASK [osism.services.traefik : Copy configuration files] ***********************",
                                "changed: [testbed-manager] => (item=traefik.yml)",
                                "changed: [testbed-manager] => (item=traefik.env)",
                                "changed: [testbed-manager] => (item=certificates.yml)",
                                "",
                                "TASK [osism.services.traefik : Copy certificate cert files] ********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy certificate key files] *********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy dynamic configuration] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Remove dynamic configuration] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create traefik external network] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Manage traefik service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.traefik : Restart traefik service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Deploy manager service] **************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply manager role] ******************************************************",
                                "included: osism.services.manager for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/ansible)",
                                "changed: [testbed-manager] => (item=/opt/archive)",
                                "changed: [testbed-manager] => (item=/opt/manager/configuration)",
                                "changed: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "changed: [testbed-manager] => (item=/opt/manager/secrets)",
                                "changed: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "changed: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "changed: [testbed-manager] => (item=ara)",
                                "changed: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "changed: [testbed-manager] => (item=conductor)",
                                "changed: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "changed: [testbed-manager] => (item=osism)",
                                "changed: [testbed-manager] => (item=osism-update-docker)",
                                "changed: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "changed: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=13   rescued=0    ignored=0   ",
                                "",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "manager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago        Up About a minute (healthy)   192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago        Up About a minute (healthy)   8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago        Up About a minute (healthy)   ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago        Up About a minute (healthy)   ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 57 seconds (healthy)       ",
                                "manager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago        Up About a minute (healthy)   ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago        Up About a minute (healthy)   3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago        Up About a minute (healthy)   ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago        Up About a minute (healthy)   6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago        Up About a minute             192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   ",
                                "osismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago        Up About a minute (healthy)   ",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg",
                                "+ osism apply resolvconf -l testbed-manager",
                                "\u001b[32m2026-01-13 00:18:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3d37815b-b49d-49a6-90f4-18332c355b3c (resolvconf) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:18:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3d37815b-b49d-49a6-90f4-18332c355b3c (resolvconf) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role resolvconf] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 13 January 2026  00:18:55 +0000 (0:00:00.144)       0:00:00.144 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Tuesday 13 January 2026  00:18:58 +0000 (0:00:03.602)       0:00:03.746 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Tuesday 13 January 2026  00:18:58 +0000 (0:00:00.067)       0:00:03.814 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Tuesday 13 January 2026  00:18:58 +0000 (0:00:00.089)       0:00:03.904 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Tuesday 13 January 2026  00:18:58 +0000 (0:00:00.073)       0:00:03.978 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Tuesday 13 January 2026  00:19:00 +0000 (0:00:01.066)       0:00:05.045 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Tuesday 13 January 2026  00:19:00 +0000 (0:00:00.063)       0:00:05.108 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Tuesday 13 January 2026  00:19:00 +0000 (0:00:00.476)       0:00:05.585 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Tuesday 13 January 2026  00:19:00 +0000 (0:00:00.075)       0:00:05.660 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Tuesday 13 January 2026  00:19:01 +0000 (0:00:00.529)       0:00:06.190 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Tuesday 13 January 2026  00:19:02 +0000 (0:00:01.060)       0:00:07.250 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Tuesday 13 January 2026  00:19:03 +0000 (0:00:00.951)       0:00:08.201 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Tuesday 13 January 2026  00:19:03 +0000 (0:00:00.086)       0:00:08.287 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:19:04 +0000 (0:00:01.139)       0:00:09.427 ******* ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.60s",
                                "osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.14s",
                                "osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 1.07s",
                                "osism.commons.resolvconf : Copy configuration files --------------------- 1.06s",
                                "osism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.95s",
                                "osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.53s",
                                "osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.48s",
                                "osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s",
                                "osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.09s",
                                "osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.08s",
                                "osism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s",
                                "osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.07s",
                                "osism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s",
                                "+ osism apply sshconfig",
                                "\u001b[32m2026-01-13 00:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 089338ef-ce47-42f2-90d9-dd2487ae976e (sshconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 089338ef-ce47-42f2-90d9-dd2487ae976e (sshconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role sshconfig] ****************************************************",
                                "",
                                "TASK [osism.commons.sshconfig : Get home directory of operator user] ***********",
                                "Tuesday 13 January 2026  00:19:20 +0000 (0:00:00.155)       0:00:00.155 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************",
                                "Tuesday 13 January 2026  00:19:21 +0000 (0:00:00.564)       0:00:00.719 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure config for each host exist] *************",
                                "Tuesday 13 January 2026  00:19:21 +0000 (0:00:00.400)       0:00:01.119 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Add extra config] ******************************",
                                "Tuesday 13 January 2026  00:19:27 +0000 (0:00:05.136)       0:00:06.256 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Assemble ssh config] ***************************",
                                "Tuesday 13 January 2026  00:19:27 +0000 (0:00:00.081)       0:00:06.337 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:19:27 +0000 (0:00:00.575)       0:00:06.912 ******* ",
                                "=============================================================================== ",
                                "osism.commons.sshconfig : Ensure config for each host exist ------------- 5.14s",
                                "osism.commons.sshconfig : Assemble ssh config --------------------------- 0.58s",
                                "osism.commons.sshconfig : Get home directory of operator user ----------- 0.56s",
                                "osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.40s",
                                "osism.commons.sshconfig : Add extra config ------------------------------ 0.08s",
                                "+ osism apply known-hosts",
                                "\u001b[32m2026-01-13 00:19:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 30ba9bc7-417f-4a1b-9d72-ca014fae44b4 (known-hosts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:19:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 30ba9bc7-417f-4a1b-9d72-ca014fae44b4 (known-hosts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role known_hosts] **************************************************",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***",
                                "Tuesday 13 January 2026  00:19:44 +0000 (0:00:00.122)       0:00:00.122 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***",
                                "Tuesday 13 January 2026  00:19:50 +0000 (0:00:05.749)       0:00:05.872 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:50 +0000 (0:00:00.161)       0:00:06.033 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEwdqUB7HN4VL2b39E+wmGAVzyJqO2pg6Y5v1/wgoBcJnDKVoLinltTeLID6TqK0rNRIrSeCdJ7ghn9+OntawF8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDhG4+2lbFna6pQJ31DRXMxGyaPNBhYP+SgYYY5Vj+SwWIJllHlv7shVTlowhDNWJH4AM5sgwNoKcSveeq58h3JoZrMC54iWMBHf7+7MpWa/kdQq9he91PMOGXYjjoaF4+l0qRkMDWTLgNz8E1qQEcH2xX89F+GlZF/3Kb8EaAnLe2ckhBlt+dydNlJPgNWT5hjgHgHCeVYe2KKhlXogk4CFVQr5YwnRO0O0vOONQ2Ryefm146Wyld/tV/22YAA7LItrYkJNGl14yYQy38A5NQUnMgHzCMMFzrPhz/IVuy8TUKyds1DrVjv1ZQjCLM1V2Xn5ENanCKlDLy1Sg774GRHwyOvoER4GADruHMhUn7DFJQo3a6vr7rUFq8gtvV6L54EFVDiuO0XqztSa4uSirBPZv0G/jM9fLDt850YS+jN2nFQATOYQvt5ADRFqUOs1MUfP02mCECk4SOakrIPRINL/gIUrFAbah16IovwWreAg4EuzwOSi0mG5l1P4wrRztk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIG1wld3KLY273VgRhq3DY4Y5VreDQGtCBE/2+XB2y63y)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:51 +0000 (0:00:01.143)       0:00:07.177 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAII3cBlxJ+Nj+vcX33DDsiEGmFzhWhUplm9v08bHaGxHd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDMNy4Ll3tlu7tXNUF3crkWBEoB1g1PQzruThglC08FOVDD6ejeWmgWy/vre8E+ErGv22sgxBVVepNWHM6ektcqb92n6EoG348bB9Zc6/dZZEEDIMvkuDQujy4Q+enMJKou3EIC8IkKUz7Qwp5nIKTTm0mOLXkmrXMYyA5TfdPaYwfncOIOmbydM05j6gD4KbIxQabeevh29UxN2KeEQTBh0Aaiw/CJTPW/sObRcdr6PiyCWPiUVAZeFSzS2ESfTVBwO0CLhtc5+lG3Z8jS0Asnu+hH6BLqzoU2I07edidZM3QiAtugKAbmdzf7f1+a88PgwxBGqT4AeBrwN4BoNVQBWYcP3fwnASx9koirDoWHqT5As8SUc2Zi/AcNUsrVpgwhxpafm4SkF+gvu3v1gFxhJ3+2bFbdrKLlDwxMM0vOvqHYAF32q9PM5vY9ZsqaDY+R2kYw65yed75mlfT+z11I7mTLiJegVMIayC9Ms6ULIyqKkpuE35qJx+zeUHS0D/M=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCANVsfM+rupUfrqcS59qXGwJqieAokpdxEwmlCCPh+A0evjKfvhj+QDlIGxgTNnid4YWZD3owkOldDVz1czbl4=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:52 +0000 (0:00:01.111)       0:00:08.288 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCs4yVojS8g9mYfYUGVbXOe9f9Q4SCvWhW3wolwzYPHIv9Xbbq3IpBpMmRarHfb4EbZ7MLtSbcxXJIWOJ5uT/v1eyT5T+8N2wVHQcZDGLxEAXU6P/wxNX1zzO9k+aHpUchtOIy01JN2dXZs5Sy5k1ZSuceJa3W3bMEdGGZ2875Lx5KDP/ivLciez4rUjtCdI0zGw/yIFfZDmiCMNBf1J4CJkuRRtEPnGhvch3AwM/2ma2fJOCFqDynv4wsIAH+5wXXeGKmtJhApo2jdhDpXlx8dDw8iE638+4gHpBb7t6EnvtfAUbp1BPdruiocm85F7TrtzepT0F8F7bCHuCCsDhZ+7Lq5VyD1GBXQTy9T4txIbj1cfYd7pP1ztsCNGtsfs2Te2qzihxL09FSWaKXe3oH0kyyjs6oAGJEPUYIa14l0Nhwf/F/Y39NcTJcpKQO5cVItDmypoSsRW/FmwWSfj4qLgjvG6laR+WwQtQ2BHQKU5OEHMaUO4nPwpOaCldDAcHk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIIb9uQuSUhF1yMMy/r7j8CSdaGWoRrCFZZtsbBuHk1lnHQp2c8Q2i0dbJwrz2NEaAHO2XIpenm+NecOa6v3p2U=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJDk/kf1I0kvvlU6paCuuF/kJCN7s38Ieq6GukgVT+0z)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:53 +0000 (0:00:01.034)       0:00:09.323 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBBT3kQanK+MPgr+KonMfs6XdQsNGu/kU8ttz39IgxJTK4uiJwzFrECUpR4sjnyGb7etmLer6xO0aafH+oD8Jc4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDSzAl+AST4boAzW56Ko76Dtd1M0bcRU4tyJFCjJH0MXVG7lRT6+RfPSA0kGrSjafm6IWd4oVyiHfYmxLxlw9qbyURve7RwylY9CnXoklCsk/mlNJ+OA6ev3zoe/1+oRcgf8zLlUgytux5eyVs0gQL0SqQ10VPq0h0zIJplqPKKhlbQQAxps722qV7ak0MWY6iet9q+AQvrnQUK9dUyhHjP12RIQaUUNtn7JJQT/wUNAk+1nNKSTrb9mnQd5Rf9ak/fxAzBHjTctwf8f5p5GURoEAwRbUikWILvHlP8adxBdS0LfmqhS0W8WyGSb70yReIKzWpIxbjf9nkEwaAYrAXsvwKSGZXz6wGtri/OvXaHpXd86ysh7zku38aMFCpkZNUPaok6j8AQ7Wwk/HgE6BM6DlVjjttr7tnmuiia7nH7El2/x06QeaTHBHD9LAyvSbfaLGr9RldoEHv3GrSeVgd6zmlsWAuJH+dMbbbii4vfL+snp5Uv+ZJYpM34r2S7Fp8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKKhUM0juZlhzLj/K6dKqkuv5KvYz7H8RaW5TIFwpjIo)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:54 +0000 (0:00:01.023)       0:00:10.346 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDNRkg+yOke3Ddbei3PVLXSl/gJWdLCOPL0+k2kVStf6W6HnoSVqAA5dQQ3nXjW8qov3yO+0olH+j2KXOlaK3tY=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiIV9KgZPNLypT6BgMveQMLb+XP2pvowAeWbYSVy5SRDT4C/PTEvxgtPE0/a2Uq1ynUXQr9qfIMI1ISnuQ2P189RXwLdzXsuECbWwQRc+iTH+7NZOnsONukNVYPuQw772ZM+l/r6CdNv0Z0dYcm9DO7MQGAWSrEZoMCeTfvACeVD+SxZNioE13o0N+riGgnwJppHHMjh5ILfJ7eu9yfKGgqOij82UBLwmWSliga7vkTObjQ5UyfGeju1sf7wsx1akOGSgbFTFRUvaQPUAJJ5G862lRN5n+iXZvlytzFDs8UFgVnRDPSRD+96KQ9c2An7hpBMe1EJHbaaHWB4btrOcRrZY2dzbN3rrVKCTuf2v/Sju6B/hexQ582byuG6WUf7Q5mw2HV0yRct3CpB8rCwKy1IKdtrjncIQ47G5XAfpDWGufIG4h14ZgdO4Gb9zp8xRi11VauLn1YL2YuTVWqgbrjeCXA7gmsrLGpa2dOV1n/ByXe53b4DQr1gq0XufLzak=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILYwppQSAKDyEbxkq0PEPVz0XhqRpbAsz1hlj5fhJOjq)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:55 +0000 (0:00:00.974)       0:00:11.320 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBiN8QzGQv8ZHPfLAudwKKlhtdteTM1C6Cw2AUpga87mGiT8xUv81/alwHnXJq3Ey3e9/R5sm9J1abxgzuI1J8M7owK/mN4T2LlXSJHlagJNjQvfKlmxZz1l+kppPMKfXcaLTwzp9gQUsMjTj4UGuO/PM4ldgmbE/NhWO6afrVT/tJnFN+UM2UgJ5BCR2N7rtPBPmI00CQGWr/WPcjDIT9aFNrZR9wr4xJwhSDePHSUAN8QjR3AEfr9ihcSBM05ESY9JhmHgFHiWyYsoCclIMclmBl+c7Bb2Sb0PlBoEH6kaM+vWS68L4i7A1EHJ/muJco6MYUofRimAYxW2+fS+Ehb6KA9Cse2F5u/b5rRcD/ZOoRgnvbl/e4M7EYCUrzOE2exqKl1Bl6hyJz4gmBejh7/Hoh0vIKqRryeuojSyAr76CegIJeSVMXHuF2l3V6qZ28UaPelyYqGbyQZdSG4S5uJOt+Cwp0neF9Egk2Sr/wvzmo0CMhzcekdV1l0uwp8CM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOpm2DX+iuhXBHOqADVnwtARbnSmUmbxHxRfUznOX5jn2moh3ojj7q8FUsWlwITOW4zsjw7yBh1DDEJH7LBlkQ4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBRKL3VzcKtOBW81QKJd742xWys9tDvetvUB2cavjZrf)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:19:56 +0000 (0:00:00.973)       0:00:12.293 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC2PZmaIQFk5fXdGfNQ+ybya9B5NctgaD8n6aoYA9hO92qMhg4UMz3L7ulQZarThvALWLN043WlfxapsN/zRnL6sK/yIXN+Em7Opi+v5NQCebcQB7xR0NB7dKLwW2ocTgjkNn8SkZ5scnUUh56VdCz+tpXorPuAGfbGPBuHjF35oiCOezymbJ4aeWj+wP/JeyE5XLWSS79QoosNLYfj2Yunul8lbPJPdSjXMXOp/DLuj/u/RM5jN0O5JU1digqgRnnQGu6TGu4le4RaDc9XozNlkrt6sN8fLwMC3oeIex0CWE1HoLS41VA6PHUlSabJ1xksQ0VpmontWHSIBuFvNAKG8gaFxIVlolm8T91B8wKx4F9hcZLgGvkhoAzancJ5itgN7xFopLQLa+D+GaFlta7DXGTTz4LllEb52vWB6U/4MIsaOJ3HDdss7MTR3vlK4mJgA0IZ8OPeitHhpeeycZPT8lumHDpN1f1W8d91App9YY4k4ZED5u052Tmh0fEV+RM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPc5npEgejWX+Bx8E+UvNX2sIZ3DqVmvOFlDMHrWgyf787GSnJk71oQxAe6VCEQGIy0fCbdGZCnpoA9V/nv1UtQ=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGOl+o72g5WDnUdyk5PWe2VS1qTxhFD/4XNOHzgBM+4b)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***",
                                "Tuesday 13 January 2026  00:19:58 +0000 (0:00:02.030)       0:00:14.324 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***",
                                "Tuesday 13 January 2026  00:20:03 +0000 (0:00:05.221)       0:00:19.546 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:03 +0000 (0:00:00.166)       0:00:19.713 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEwdqUB7HN4VL2b39E+wmGAVzyJqO2pg6Y5v1/wgoBcJnDKVoLinltTeLID6TqK0rNRIrSeCdJ7ghn9+OntawF8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDhG4+2lbFna6pQJ31DRXMxGyaPNBhYP+SgYYY5Vj+SwWIJllHlv7shVTlowhDNWJH4AM5sgwNoKcSveeq58h3JoZrMC54iWMBHf7+7MpWa/kdQq9he91PMOGXYjjoaF4+l0qRkMDWTLgNz8E1qQEcH2xX89F+GlZF/3Kb8EaAnLe2ckhBlt+dydNlJPgNWT5hjgHgHCeVYe2KKhlXogk4CFVQr5YwnRO0O0vOONQ2Ryefm146Wyld/tV/22YAA7LItrYkJNGl14yYQy38A5NQUnMgHzCMMFzrPhz/IVuy8TUKyds1DrVjv1ZQjCLM1V2Xn5ENanCKlDLy1Sg774GRHwyOvoER4GADruHMhUn7DFJQo3a6vr7rUFq8gtvV6L54EFVDiuO0XqztSa4uSirBPZv0G/jM9fLDt850YS+jN2nFQATOYQvt5ADRFqUOs1MUfP02mCECk4SOakrIPRINL/gIUrFAbah16IovwWreAg4EuzwOSi0mG5l1P4wrRztk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIG1wld3KLY273VgRhq3DY4Y5VreDQGtCBE/2+XB2y63y)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:04 +0000 (0:00:01.018)       0:00:20.731 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAII3cBlxJ+Nj+vcX33DDsiEGmFzhWhUplm9v08bHaGxHd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDMNy4Ll3tlu7tXNUF3crkWBEoB1g1PQzruThglC08FOVDD6ejeWmgWy/vre8E+ErGv22sgxBVVepNWHM6ektcqb92n6EoG348bB9Zc6/dZZEEDIMvkuDQujy4Q+enMJKou3EIC8IkKUz7Qwp5nIKTTm0mOLXkmrXMYyA5TfdPaYwfncOIOmbydM05j6gD4KbIxQabeevh29UxN2KeEQTBh0Aaiw/CJTPW/sObRcdr6PiyCWPiUVAZeFSzS2ESfTVBwO0CLhtc5+lG3Z8jS0Asnu+hH6BLqzoU2I07edidZM3QiAtugKAbmdzf7f1+a88PgwxBGqT4AeBrwN4BoNVQBWYcP3fwnASx9koirDoWHqT5As8SUc2Zi/AcNUsrVpgwhxpafm4SkF+gvu3v1gFxhJ3+2bFbdrKLlDwxMM0vOvqHYAF32q9PM5vY9ZsqaDY+R2kYw65yed75mlfT+z11I7mTLiJegVMIayC9Ms6ULIyqKkpuE35qJx+zeUHS0D/M=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCANVsfM+rupUfrqcS59qXGwJqieAokpdxEwmlCCPh+A0evjKfvhj+QDlIGxgTNnid4YWZD3owkOldDVz1czbl4=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:05 +0000 (0:00:01.026)       0:00:21.758 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCs4yVojS8g9mYfYUGVbXOe9f9Q4SCvWhW3wolwzYPHIv9Xbbq3IpBpMmRarHfb4EbZ7MLtSbcxXJIWOJ5uT/v1eyT5T+8N2wVHQcZDGLxEAXU6P/wxNX1zzO9k+aHpUchtOIy01JN2dXZs5Sy5k1ZSuceJa3W3bMEdGGZ2875Lx5KDP/ivLciez4rUjtCdI0zGw/yIFfZDmiCMNBf1J4CJkuRRtEPnGhvch3AwM/2ma2fJOCFqDynv4wsIAH+5wXXeGKmtJhApo2jdhDpXlx8dDw8iE638+4gHpBb7t6EnvtfAUbp1BPdruiocm85F7TrtzepT0F8F7bCHuCCsDhZ+7Lq5VyD1GBXQTy9T4txIbj1cfYd7pP1ztsCNGtsfs2Te2qzihxL09FSWaKXe3oH0kyyjs6oAGJEPUYIa14l0Nhwf/F/Y39NcTJcpKQO5cVItDmypoSsRW/FmwWSfj4qLgjvG6laR+WwQtQ2BHQKU5OEHMaUO4nPwpOaCldDAcHk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIIb9uQuSUhF1yMMy/r7j8CSdaGWoRrCFZZtsbBuHk1lnHQp2c8Q2i0dbJwrz2NEaAHO2XIpenm+NecOa6v3p2U=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJDk/kf1I0kvvlU6paCuuF/kJCN7s38Ieq6GukgVT+0z)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:07 +0000 (0:00:01.050)       0:00:22.808 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBBT3kQanK+MPgr+KonMfs6XdQsNGu/kU8ttz39IgxJTK4uiJwzFrECUpR4sjnyGb7etmLer6xO0aafH+oD8Jc4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDSzAl+AST4boAzW56Ko76Dtd1M0bcRU4tyJFCjJH0MXVG7lRT6+RfPSA0kGrSjafm6IWd4oVyiHfYmxLxlw9qbyURve7RwylY9CnXoklCsk/mlNJ+OA6ev3zoe/1+oRcgf8zLlUgytux5eyVs0gQL0SqQ10VPq0h0zIJplqPKKhlbQQAxps722qV7ak0MWY6iet9q+AQvrnQUK9dUyhHjP12RIQaUUNtn7JJQT/wUNAk+1nNKSTrb9mnQd5Rf9ak/fxAzBHjTctwf8f5p5GURoEAwRbUikWILvHlP8adxBdS0LfmqhS0W8WyGSb70yReIKzWpIxbjf9nkEwaAYrAXsvwKSGZXz6wGtri/OvXaHpXd86ysh7zku38aMFCpkZNUPaok6j8AQ7Wwk/HgE6BM6DlVjjttr7tnmuiia7nH7El2/x06QeaTHBHD9LAyvSbfaLGr9RldoEHv3GrSeVgd6zmlsWAuJH+dMbbbii4vfL+snp5Uv+ZJYpM34r2S7Fp8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKKhUM0juZlhzLj/K6dKqkuv5KvYz7H8RaW5TIFwpjIo)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:08 +0000 (0:00:01.046)       0:00:23.855 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDNRkg+yOke3Ddbei3PVLXSl/gJWdLCOPL0+k2kVStf6W6HnoSVqAA5dQQ3nXjW8qov3yO+0olH+j2KXOlaK3tY=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCiIV9KgZPNLypT6BgMveQMLb+XP2pvowAeWbYSVy5SRDT4C/PTEvxgtPE0/a2Uq1ynUXQr9qfIMI1ISnuQ2P189RXwLdzXsuECbWwQRc+iTH+7NZOnsONukNVYPuQw772ZM+l/r6CdNv0Z0dYcm9DO7MQGAWSrEZoMCeTfvACeVD+SxZNioE13o0N+riGgnwJppHHMjh5ILfJ7eu9yfKGgqOij82UBLwmWSliga7vkTObjQ5UyfGeju1sf7wsx1akOGSgbFTFRUvaQPUAJJ5G862lRN5n+iXZvlytzFDs8UFgVnRDPSRD+96KQ9c2An7hpBMe1EJHbaaHWB4btrOcRrZY2dzbN3rrVKCTuf2v/Sju6B/hexQ582byuG6WUf7Q5mw2HV0yRct3CpB8rCwKy1IKdtrjncIQ47G5XAfpDWGufIG4h14ZgdO4Gb9zp8xRi11VauLn1YL2YuTVWqgbrjeCXA7gmsrLGpa2dOV1n/ByXe53b4DQr1gq0XufLzak=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILYwppQSAKDyEbxkq0PEPVz0XhqRpbAsz1hlj5fhJOjq)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:09 +0000 (0:00:01.059)       0:00:24.915 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBRKL3VzcKtOBW81QKJd742xWys9tDvetvUB2cavjZrf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBiN8QzGQv8ZHPfLAudwKKlhtdteTM1C6Cw2AUpga87mGiT8xUv81/alwHnXJq3Ey3e9/R5sm9J1abxgzuI1J8M7owK/mN4T2LlXSJHlagJNjQvfKlmxZz1l+kppPMKfXcaLTwzp9gQUsMjTj4UGuO/PM4ldgmbE/NhWO6afrVT/tJnFN+UM2UgJ5BCR2N7rtPBPmI00CQGWr/WPcjDIT9aFNrZR9wr4xJwhSDePHSUAN8QjR3AEfr9ihcSBM05ESY9JhmHgFHiWyYsoCclIMclmBl+c7Bb2Sb0PlBoEH6kaM+vWS68L4i7A1EHJ/muJco6MYUofRimAYxW2+fS+Ehb6KA9Cse2F5u/b5rRcD/ZOoRgnvbl/e4M7EYCUrzOE2exqKl1Bl6hyJz4gmBejh7/Hoh0vIKqRryeuojSyAr76CegIJeSVMXHuF2l3V6qZ28UaPelyYqGbyQZdSG4S5uJOt+Cwp0neF9Egk2Sr/wvzmo0CMhzcekdV1l0uwp8CM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOpm2DX+iuhXBHOqADVnwtARbnSmUmbxHxRfUznOX5jn2moh3ojj7q8FUsWlwITOW4zsjw7yBh1DDEJH7LBlkQ4=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 13 January 2026  00:20:10 +0000 (0:00:01.040)       0:00:25.955 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC2PZmaIQFk5fXdGfNQ+ybya9B5NctgaD8n6aoYA9hO92qMhg4UMz3L7ulQZarThvALWLN043WlfxapsN/zRnL6sK/yIXN+Em7Opi+v5NQCebcQB7xR0NB7dKLwW2ocTgjkNn8SkZ5scnUUh56VdCz+tpXorPuAGfbGPBuHjF35oiCOezymbJ4aeWj+wP/JeyE5XLWSS79QoosNLYfj2Yunul8lbPJPdSjXMXOp/DLuj/u/RM5jN0O5JU1digqgRnnQGu6TGu4le4RaDc9XozNlkrt6sN8fLwMC3oeIex0CWE1HoLS41VA6PHUlSabJ1xksQ0VpmontWHSIBuFvNAKG8gaFxIVlolm8T91B8wKx4F9hcZLgGvkhoAzancJ5itgN7xFopLQLa+D+GaFlta7DXGTTz4LllEb52vWB6U/4MIsaOJ3HDdss7MTR3vlK4mJgA0IZ8OPeitHhpeeycZPT8lumHDpN1f1W8d91App9YY4k4ZED5u052Tmh0fEV+RM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGOl+o72g5WDnUdyk5PWe2VS1qTxhFD/4XNOHzgBM+4b)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPc5npEgejWX+Bx8E+UvNX2sIZ3DqVmvOFlDMHrWgyf787GSnJk71oQxAe6VCEQGIy0fCbdGZCnpoA9V/nv1UtQ=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write static known_hosts entries] ************",
                                "Tuesday 13 January 2026  00:20:11 +0000 (0:00:01.083)       0:00:27.038 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write extra known_hosts entries] *************",
                                "Tuesday 13 January 2026  00:20:11 +0000 (0:00:00.156)       0:00:27.195 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Delete known_hosts entries] ******************",
                                "Tuesday 13 January 2026  00:20:11 +0000 (0:00:00.066)       0:00:27.261 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Set file permissions] ************************",
                                "Tuesday 13 January 2026  00:20:11 +0000 (0:00:00.055)       0:00:27.317 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:20:12 +0000 (0:00:00.703)       0:00:28.020 ******* ",
                                "=============================================================================== ",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.75s",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.22s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 2.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.14s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.11s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.08s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.06s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s",
                                "osism.commons.known_hosts : Set file permissions ------------------------ 0.70s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s",
                                "osism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s",
                                "+ osism apply squid",
                                "\u001b[32m2026-01-13 00:20:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e69aa7d1-79c0-4f37-a147-7b4d8d64571b (squid) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:20:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task e69aa7d1-79c0-4f37-a147-7b4d8d64571b (squid) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role squid] ********************************************************",
                                "",
                                "TASK [osism.services.squid : Include install tasks] ****************************",
                                "Tuesday 13 January 2026  00:20:28 +0000 (0:00:00.174)       0:00:00.174 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Install required packages] ************************",
                                "Tuesday 13 January 2026  00:20:28 +0000 (0:00:00.085)       0:00:00.260 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Create required directories] **********************",
                                "Tuesday 13 January 2026  00:20:30 +0000 (0:00:01.433)       0:00:01.694 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy squid configuration files] *******************",
                                "Tuesday 13 January 2026  00:20:31 +0000 (0:00:01.121)       0:00:02.816 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***",
                                "Tuesday 13 January 2026  00:20:32 +0000 (0:00:01.093)       0:00:03.910 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy docker-compose.yml file] *********************",
                                "Tuesday 13 January 2026  00:20:32 +0000 (0:00:00.316)       0:00:04.226 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Manage squid service] *****************************",
                                "Tuesday 13 January 2026  00:20:33 +0000 (0:00:00.901)       0:00:05.128 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Restart squid service] *****************",
                                "Tuesday 13 January 2026  00:21:04 +0000 (0:00:30.388)       0:00:35.516 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******",
                                "Tuesday 13 January 2026  00:21:19 +0000 (0:00:15.719)       0:00:51.236 ******* ",
                                "Pausing for 60 seconds",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***",
                                "Tuesday 13 January 2026  00:22:20 +0000 (0:01:00.081)       0:01:51.317 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****",
                                "Tuesday 13 January 2026  00:22:20 +0000 (0:00:00.067)       0:01:51.385 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:22:20 +0000 (0:00:00.568)       0:01:51.954 ******* ",
                                "=============================================================================== ",
                                "osism.services.squid : Wait for squid service to start ----------------- 60.08s",
                                "osism.services.squid : Manage squid service ---------------------------- 30.39s",
                                "osism.services.squid : Restart squid service --------------------------- 15.72s",
                                "osism.services.squid : Install required packages ------------------------ 1.43s",
                                "osism.services.squid : Create required directories ---------------------- 1.12s",
                                "osism.services.squid : Copy squid configuration files ------------------- 1.09s",
                                "osism.services.squid : Copy docker-compose.yml file --------------------- 0.90s",
                                "osism.services.squid : Wait for an healthy squid service ---------------- 0.57s",
                                "osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.32s",
                                "osism.services.squid : Include install tasks ---------------------------- 0.09s",
                                "osism.services.squid : Register that squid service was restarted -------- 0.07s",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla",
                                "+ set -e",
                                "+ NAMESPACE=kolla",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "++ semver latest 9.0.0",
                                "+ [[ -1 -lt 0 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply operator -u ubuntu -l testbed-nodes",
                                "\u001b[32m2026-01-13 00:22:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06e949a0-1f23-46bf-b910-fb634889464f (operator) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:22:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06e949a0-1f23-46bf-b910-fb634889464f (operator) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Make ssh pipelining working] *********************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 13 January 2026  00:22:37 +0000 (0:00:00.135)       0:00:00.135 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Do not require tty for all users] ****************************************",
                                "Tuesday 13 January 2026  00:22:41 +0000 (0:00:04.319)       0:00:04.455 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role operator] *****************************************************",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "Tuesday 13 January 2026  00:22:42 +0000 (0:00:00.766)       0:00:05.221 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "Tuesday 13 January 2026  00:22:42 +0000 (0:00:00.184)       0:00:05.405 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "Tuesday 13 January 2026  00:22:42 +0000 (0:00:00.170)       0:00:05.576 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "Tuesday 13 January 2026  00:22:43 +0000 (0:00:00.631)       0:00:06.208 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "Tuesday 13 January 2026  00:22:44 +0000 (0:00:00.749)       0:00:06.957 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "Tuesday 13 January 2026  00:22:45 +0000 (0:00:01.220)       0:00:08.177 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "Tuesday 13 January 2026  00:22:46 +0000 (0:00:01.165)       0:00:09.343 ******* ",
                                "\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m",
                                "\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m",
                                "\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "Tuesday 13 January 2026  00:22:47 +0000 (0:00:01.204)       0:00:10.547 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "Tuesday 13 January 2026  00:22:47 +0000 (0:00:00.152)       0:00:10.699 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "Tuesday 13 January 2026  00:22:48 +0000 (0:00:00.169)       0:00:10.869 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "Tuesday 13 January 2026  00:22:48 +0000 (0:00:00.592)       0:00:11.461 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "Tuesday 13 January 2026  00:22:48 +0000 (0:00:00.151)       0:00:11.613 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "Tuesday 13 January 2026  00:22:49 +0000 (0:00:00.686)       0:00:12.300 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "Tuesday 13 January 2026  00:22:49 +0000 (0:00:00.152)       0:00:12.452 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "Tuesday 13 January 2026  00:22:49 +0000 (0:00:00.150)       0:00:12.603 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "Tuesday 13 January 2026  00:22:49 +0000 (0:00:00.152)       0:00:12.756 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "Tuesday 13 January 2026  00:22:50 +0000 (0:00:00.663)       0:00:13.419 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:22:50 +0000 (0:00:00.262)       0:00:13.682 ******* ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 4.32s",
                                "osism.commons.operator : Add user to additional groups ------------------ 1.22s",
                                "osism.commons.operator : Set language variables in .bashrc configuration file --- 1.20s",
                                "osism.commons.operator : Copy user sudoers file ------------------------- 1.17s",
                                "Do not require tty for all users ---------------------------------------- 0.77s",
                                "osism.commons.operator : Create user ------------------------------------ 0.75s",
                                "osism.commons.operator : Set ssh authorized keys ------------------------ 0.69s",
                                "osism.commons.operator : Set password ----------------------------------- 0.66s",
                                "osism.commons.operator : Create operator group -------------------------- 0.63s",
                                "osism.commons.operator : Create .ssh directory -------------------------- 0.59s",
                                "osism.commons.operator : Unset & lock password -------------------------- 0.26s",
                                "osism.commons.operator : Gather variables for each operating system ----- 0.18s",
                                "osism.commons.operator : Set operator_groups variable to default value --- 0.17s",
                                "osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.17s",
                                "osism.commons.operator : Delete authorized GitHub accounts -------------- 0.15s",
                                "osism.commons.operator : Delete ssh authorized keys --------------------- 0.15s",
                                "osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.15s",
                                "osism.commons.operator : Check number of SSH authorized keys ------------ 0.15s",
                                "osism.commons.operator : Set authorized GitHub accounts ----------------- 0.15s",
                                "+ osism apply --environment custom facts",
                                "\u001b[32m2026-01-13 00:22:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m",
                                "\u001b[32m2026-01-13 00:23:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 680fbef7-b83d-4dcb-b9b6-a0e0d6048bcb (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:23:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 680fbef7-b83d-4dcb-b9b6-a0e0d6048bcb (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy custom network devices fact] ****************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Tuesday 13 January 2026  00:23:07 +0000 (0:00:00.073)       0:00:00.073 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Copy fact file] **********************************************************",
                                "Tuesday 13 January 2026  00:23:08 +0000 (0:00:01.419)       0:00:01.493 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Copy custom ceph devices facts] ******************************************",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Tuesday 13 January 2026  00:23:09 +0000 (0:00:01.089)       0:00:02.583 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Tuesday 13 January 2026  00:23:10 +0000 (0:00:00.078)       0:00:02.661 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Tuesday 13 January 2026  00:23:10 +0000 (0:00:00.182)       0:00:02.844 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Tuesday 13 January 2026  00:23:10 +0000 (0:00:00.183)       0:00:03.027 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Tuesday 13 January 2026  00:23:10 +0000 (0:00:00.127)       0:00:03.155 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Tuesday 13 January 2026  00:23:10 +0000 (0:00:00.424)       0:00:03.579 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Tuesday 13 January 2026  00:23:11 +0000 (0:00:00.112)       0:00:03.692 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Tuesday 13 January 2026  00:23:12 +0000 (0:00:00.997)       0:00:04.689 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Tuesday 13 January 2026  00:23:12 +0000 (0:00:00.464)       0:00:05.154 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Tuesday 13 January 2026  00:23:13 +0000 (0:00:01.046)       0:00:06.200 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Install required packages (RedHat)] **************************************",
                                "Tuesday 13 January 2026  00:23:29 +0000 (0:00:15.507)       0:00:21.708 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (Debian)] **************************************",
                                "Tuesday 13 January 2026  00:23:29 +0000 (0:00:00.100)       0:00:21.808 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Tuesday 13 January 2026  00:23:37 +0000 (0:00:08.343)       0:00:30.152 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "Tuesday 13 January 2026  00:23:37 +0000 (0:00:00.458)       0:00:30.610 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Tuesday 13 January 2026  00:23:41 +0000 (0:00:03.568)       0:00:34.179 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:23:42 +0000 (0:00:01.266)       0:00:35.446 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:23:46 +0000 (0:00:03.759)       0:00:39.205 ******* ",
                                "=============================================================================== ",
                                "osism.commons.repository : Update package cache ------------------------ 15.51s",
                                "Install required packages (Debian) -------------------------------------- 8.34s",
                                "Gathers facts about hosts ----------------------------------------------- 3.76s",
                                "Copy fact files --------------------------------------------------------- 3.57s",
                                "Create custom facts directory ------------------------------------------- 1.42s",
                                "osism.commons.repository : Force update of package cache ---------------- 1.27s",
                                "Copy fact file ---------------------------------------------------------- 1.09s",
                                "osism.commons.repository : Copy ubuntu.sources file --------------------- 1.05s",
                                "osism.commons.repository : Copy 99osism apt configuration --------------- 1.00s",
                                "osism.commons.repository : Remove sources.list file --------------------- 0.46s",
                                "Create custom facts directory ------------------------------------------- 0.46s",
                                "osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.42s",
                                "osism.commons.repository : Set repositories to default ------------------ 0.18s",
                                "osism.commons.repository : Set repository_default fact to default value --- 0.18s",
                                "osism.commons.repository : Include distribution specific repository tasks --- 0.13s",
                                "osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.11s",
                                "Install required packages (RedHat) -------------------------------------- 0.10s",
                                "osism.commons.repository : Gather variables for each operating system --- 0.08s",
                                "+ osism apply bootstrap",
                                "\u001b[32m2026-01-13 00:23:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b55d8260-af86-4169-9c84-5a2c3e1c880b (bootstrap) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:23:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b55d8260-af86-4169-9c84-5a2c3e1c880b (bootstrap) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on state bootstrap] ************************************",
                                "",
                                "TASK [Group hosts based on state bootstrap] ************************************",
                                "Tuesday 13 January 2026  00:24:03 +0000 (0:00:00.114)       0:00:00.114 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:24:03 +0000 (0:00:00.197)       0:00:00.312 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts (if using --limit)] ***************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:24:08 +0000 (0:00:04.697)       0:00:05.009 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 1] ********************************************",
                                "",
                                "TASK [osism.commons.hostname : Set hostname] ***********************************",
                                "Tuesday 13 January 2026  00:24:08 +0000 (0:00:00.454)       0:00:05.464 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.hostname : Copy /etc/hostname] *****************************",
                                "Tuesday 13 January 2026  00:24:10 +0000 (0:00:01.345)       0:00:06.809 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Include type specific tasks] ***********************",
                                "Tuesday 13 January 2026  00:24:11 +0000 (0:00:01.230)       0:00:08.039 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Copy /etc/hosts file] ******************************",
                                "Tuesday 13 January 2026  00:24:11 +0000 (0:00:00.263)       0:00:08.303 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Include distribution specific tasks] ***************",
                                "Tuesday 13 January 2026  00:24:13 +0000 (0:00:02.092)       0:00:10.395 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Configure proxy parameters for apt] ****************",
                                "Tuesday 13 January 2026  00:24:13 +0000 (0:00:00.250)       0:00:10.646 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Set system wide settings in environment file] ******",
                                "Tuesday 13 January 2026  00:24:15 +0000 (0:00:01.028)       0:00:11.674 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Remove system wide settings in environment file] ***",
                                "Tuesday 13 January 2026  00:24:15 +0000 (0:00:00.575)       0:00:12.250 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Tuesday 13 January 2026  00:24:15 +0000 (0:00:00.424)       0:00:12.674 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Tuesday 13 January 2026  00:24:16 +0000 (0:00:00.225)       0:00:12.900 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Tuesday 13 January 2026  00:24:16 +0000 (0:00:00.281)       0:00:13.182 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Tuesday 13 January 2026  00:24:16 +0000 (0:00:00.297)       0:00:13.479 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Tuesday 13 January 2026  00:24:18 +0000 (0:00:01.637)       0:00:15.117 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Tuesday 13 January 2026  00:24:18 +0000 (0:00:00.245)       0:00:15.363 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Tuesday 13 January 2026  00:24:19 +0000 (0:00:00.552)       0:00:15.915 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Tuesday 13 January 2026  00:24:19 +0000 (0:00:00.334)       0:00:16.250 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Tuesday 13 January 2026  00:24:20 +0000 (0:00:00.539)       0:00:16.791 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Tuesday 13 January 2026  00:24:21 +0000 (0:00:01.146)       0:00:17.937 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Tuesday 13 January 2026  00:24:22 +0000 (0:00:01.170)       0:00:19.108 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Tuesday 13 January 2026  00:24:22 +0000 (0:00:00.319)       0:00:19.427 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Tuesday 13 January 2026  00:24:24 +0000 (0:00:01.272)       0:00:20.700 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Tuesday 13 January 2026  00:24:24 +0000 (0:00:00.228)       0:00:20.928 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Tuesday 13 January 2026  00:24:24 +0000 (0:00:00.233)       0:00:21.161 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Tuesday 13 January 2026  00:24:24 +0000 (0:00:00.228)       0:00:21.390 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Tuesday 13 January 2026  00:24:25 +0000 (0:00:00.290)       0:00:21.681 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Tuesday 13 January 2026  00:24:25 +0000 (0:00:00.503)       0:00:22.185 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Tuesday 13 January 2026  00:24:25 +0000 (0:00:00.253)       0:00:22.438 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Tuesday 13 January 2026  00:24:26 +0000 (0:00:00.965)       0:00:23.403 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Tuesday 13 January 2026  00:24:27 +0000 (0:00:00.523)       0:00:23.927 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Tuesday 13 January 2026  00:24:28 +0000 (0:00:01.157)       0:00:25.084 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Gather variables for each operating system] *****",
                                "Tuesday 13 January 2026  00:24:44 +0000 (0:00:16.143)       0:00:41.228 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****",
                                "Tuesday 13 January 2026  00:24:44 +0000 (0:00:00.240)       0:00:41.468 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***",
                                "Tuesday 13 January 2026  00:24:45 +0000 (0:00:00.232)       0:00:41.701 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include distribution specific install tasks] ****",
                                "Tuesday 13 January 2026  00:24:45 +0000 (0:00:00.217)       0:00:41.919 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Install rsyslog package] ************************",
                                "Tuesday 13 January 2026  00:24:45 +0000 (0:00:00.263)       0:00:42.182 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********",
                                "Tuesday 13 January 2026  00:24:47 +0000 (0:00:01.870)       0:00:44.053 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Manage rsyslog service] *************************",
                                "Tuesday 13 January 2026  00:24:48 +0000 (0:00:01.189)       0:00:45.243 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include fluentd tasks] **************************",
                                "Tuesday 13 January 2026  00:24:49 +0000 (0:00:00.869)       0:00:46.112 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***",
                                "Tuesday 13 January 2026  00:24:49 +0000 (0:00:00.286)       0:00:46.399 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include additional log server tasks] ************",
                                "Tuesday 13 January 2026  00:24:50 +0000 (0:00:01.106)       0:00:47.505 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include logrotate tasks] ************************",
                                "Tuesday 13 January 2026  00:24:51 +0000 (0:00:00.240)       0:00:47.746 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Ensure logrotate package is installed] **********",
                                "Tuesday 13 January 2026  00:24:51 +0000 (0:00:00.292)       0:00:48.039 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************",
                                "Tuesday 13 January 2026  00:24:53 +0000 (0:00:01.825)       0:00:49.865 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Install util-linux-extra package] ****************",
                                "Tuesday 13 January 2026  00:24:54 +0000 (0:00:01.116)       0:00:50.981 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Sync hardware clock] *****************************",
                                "Tuesday 13 January 2026  00:25:06 +0000 (0:00:11.698)       0:01:02.679 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.configfs : Start sys-kernel-config mount] ******************",
                                "Tuesday 13 January 2026  00:25:07 +0000 (0:00:01.656)       0:01:04.335 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Gather variables for each operating system] *****",
                                "Tuesday 13 January 2026  00:25:08 +0000 (0:00:00.887)       0:01:05.223 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***",
                                "Tuesday 13 January 2026  00:25:08 +0000 (0:00:00.210)       0:01:05.434 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Include distribution specific package tasks] ****",
                                "Tuesday 13 January 2026  00:25:08 +0000 (0:00:00.217)       0:01:05.651 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install needrestart package] ********************",
                                "Tuesday 13 January 2026  00:25:09 +0000 (0:00:00.266)       0:01:05.917 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set needrestart mode] ***************************",
                                "Tuesday 13 January 2026  00:25:10 +0000 (0:00:01.667)       0:01:07.585 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***",
                                "Tuesday 13 January 2026  00:25:11 +0000 (0:00:00.587)       0:01:08.173 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Update package cache] ***************************",
                                "Tuesday 13 January 2026  00:25:11 +0000 (0:00:00.234)       0:01:08.408 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download upgrade packages] **********************",
                                "Tuesday 13 January 2026  00:25:13 +0000 (0:00:01.314)       0:01:09.722 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Upgrade packages] *******************************",
                                "Tuesday 13 January 2026  00:25:14 +0000 (0:00:01.735)       0:01:11.458 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download required packages] *********************",
                                "Tuesday 13 January 2026  00:25:17 +0000 (0:00:02.752)       0:01:14.210 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install required packages] **********************",
                                "Tuesday 13 January 2026  00:25:55 +0000 (0:00:38.022)       0:01:52.233 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove useless packages from the cache] *********",
                                "Tuesday 13 January 2026  00:27:12 +0000 (0:01:17.350)       0:03:09.583 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove dependencies that are no longer required] ***",
                                "Tuesday 13 January 2026  00:27:14 +0000 (0:00:01.891)       0:03:11.475 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Include sysctl tasks] *****************************",
                                "Tuesday 13 January 2026  00:27:24 +0000 (0:00:10.154)       0:03:21.630 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********",
                                "Tuesday 13 January 2026  00:27:25 +0000 (0:00:00.362)       0:03:21.992 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************",
                                "Tuesday 13 January 2026  00:27:25 +0000 (0:00:00.608)       0:03:22.601 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************",
                                "Tuesday 13 January 2026  00:27:33 +0000 (0:00:07.949)       0:03:30.550 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************",
                                "Tuesday 13 January 2026  00:27:34 +0000 (0:00:00.637)       0:03:31.188 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on network] *****************",
                                "Tuesday 13 January 2026  00:27:35 +0000 (0:00:00.617)       0:03:31.806 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************",
                                "Tuesday 13 January 2026  00:27:36 +0000 (0:00:01.604)       0:03:33.410 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "",
                                "TASK [osism.commons.limits : Include limits tasks] *****************************",
                                "Tuesday 13 January 2026  00:27:37 +0000 (0:00:00.602)       0:03:34.012 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Populate service facts] *************************",
                                "Tuesday 13 January 2026  00:27:37 +0000 (0:00:00.327)       0:03:34.340 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Check services] *********************************",
                                "Tuesday 13 January 2026  00:27:43 +0000 (0:00:05.539)       0:03:39.879 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Start/enable required services] *****************",
                                "Tuesday 13 January 2026  00:27:43 +0000 (0:00:00.328)       0:03:40.208 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Include distribution specific configure tasks] ******",
                                "Tuesday 13 January 2026  00:27:44 +0000 (0:00:01.292)       0:03:41.501 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove update-motd package] *************************",
                                "Tuesday 13 January 2026  00:27:45 +0000 (0:00:00.517)       0:03:42.018 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************",
                                "Tuesday 13 January 2026  00:27:46 +0000 (0:00:01.265)       0:03:43.284 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Disable the dynamic motd-news service] **************",
                                "Tuesday 13 January 2026  00:27:47 +0000 (0:00:00.673)       0:03:43.957 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********",
                                "Tuesday 13 January 2026  00:27:47 +0000 (0:00:00.636)       0:03:44.594 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove pam_motd.so rule] ****************************",
                                "Tuesday 13 January 2026  00:27:48 +0000 (0:00:00.666)       0:03:45.261 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262699.1741004, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262720.6874278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262712.4412637, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262727.7672517, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262721.3498943, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262703.8333607, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1768262705.3500478, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy motd file] *************************************",
                                "Tuesday 13 January 2026  00:27:49 +0000 (0:00:01.020)       0:03:46.281 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue file] ************************************",
                                "Tuesday 13 January 2026  00:27:50 +0000 (0:00:01.256)       0:03:47.538 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue.net file] ********************************",
                                "Tuesday 13 January 2026  00:27:52 +0000 (0:00:01.155)       0:03:48.693 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to print the motd] ********************",
                                "Tuesday 13 January 2026  00:27:53 +0000 (0:00:01.244)       0:03:49.938 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to not print the motd] ****************",
                                "Tuesday 13 January 2026  00:27:53 +0000 (0:00:00.290)       0:03:50.228 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Include distribution specific install tasks] ********",
                                "Tuesday 13 January 2026  00:27:54 +0000 (0:00:00.728)       0:03:50.957 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Install rng package] ********************************",
                                "Tuesday 13 January 2026  00:27:54 +0000 (0:00:00.419)       0:03:51.377 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Remove haveged package] *****************************",
                                "Tuesday 13 January 2026  00:28:03 +0000 (0:00:08.546)       0:03:59.923 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Manage rng service] *********************************",
                                "Tuesday 13 January 2026  00:28:04 +0000 (0:00:01.511)       0:04:01.434 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Gather variables for each operating system] ******",
                                "Tuesday 13 January 2026  00:28:05 +0000 (0:00:01.231)       0:04:02.665 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***",
                                "Tuesday 13 January 2026  00:28:06 +0000 (0:00:00.324)       0:04:02.990 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***",
                                "Tuesday 13 January 2026  00:28:06 +0000 (0:00:00.347)       0:04:03.338 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Populate service facts] **************************",
                                "Tuesday 13 January 2026  00:28:06 +0000 (0:00:00.318)       0:04:03.657 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include distribution specific timer tasks] *******",
                                "Tuesday 13 January 2026  00:28:11 +0000 (0:00:04.886)       0:04:08.543 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Disable apt-daily timers] ************************",
                                "Tuesday 13 January 2026  00:28:12 +0000 (0:00:00.464)       0:04:09.008 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include service tasks] ***************************",
                                "Tuesday 13 January 2026  00:28:12 +0000 (0:00:00.357)       0:04:09.365 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup services] ********************************",
                                "Tuesday 13 January 2026  00:28:13 +0000 (0:00:00.407)       0:04:09.773 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include packages tasks] **************************",
                                "Tuesday 13 January 2026  00:28:13 +0000 (0:00:00.331)       0:04:10.105 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup installed packages] **********************",
                                "Tuesday 13 January 2026  00:28:13 +0000 (0:00:00.401)       0:04:10.507 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloudinit package] ************************",
                                "Tuesday 13 January 2026  00:28:48 +0000 (0:00:34.302)       0:04:44.809 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********",
                                "Tuesday 13 January 2026  00:28:56 +0000 (0:00:08.842)       0:04:53.652 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove useless packages from the cache] **********",
                                "Tuesday 13 January 2026  00:29:05 +0000 (0:00:08.171)       0:05:01.823 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***",
                                "Tuesday 13 January 2026  00:29:07 +0000 (0:00:01.923)       0:05:03.747 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include cloudinit tasks] *************************",
                                "Tuesday 13 January 2026  00:29:13 +0000 (0:00:06.615)       0:05:10.362 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******",
                                "Tuesday 13 January 2026  00:29:14 +0000 (0:00:00.620)       0:05:10.983 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Install tzdata package] *************************",
                                "Tuesday 13 January 2026  00:29:15 +0000 (0:00:00.822)       0:05:11.805 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Set timezone to UTC] ****************************",
                                "Tuesday 13 January 2026  00:29:16 +0000 (0:00:01.768)       0:05:13.574 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Create /etc/adjtime file] ***********************",
                                "Tuesday 13 January 2026  00:29:17 +0000 (0:00:00.810)       0:05:14.385 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************",
                                "Tuesday 13 January 2026  00:29:18 +0000 (0:00:00.295)       0:05:14.680 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather variables for each operating system] ******",
                                "Tuesday 13 January 2026  00:29:18 +0000 (0:00:00.401)       0:05:15.081 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_version variable to default value] ****",
                                "Tuesday 13 January 2026  00:29:18 +0000 (0:00:00.280)       0:05:15.362 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_cli_version variable to default value] ***",
                                "Tuesday 13 January 2026  00:29:18 +0000 (0:00:00.269)       0:05:15.631 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker version] ***********************",
                                "Tuesday 13 January 2026  00:29:19 +0000 (0:00:00.280)       0:05:15.912 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker cli version] *******************",
                                "Tuesday 13 January 2026  00:29:19 +0000 (0:00:00.290)       0:05:16.203 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include block storage tasks] *********************",
                                "Tuesday 13 January 2026  00:29:19 +0000 (0:00:00.292)       0:05:16.495 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include zram storage tasks] **********************",
                                "Tuesday 13 January 2026  00:29:20 +0000 (0:00:00.267)       0:05:16.763 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include docker install tasks] ********************",
                                "Tuesday 13 January 2026  00:29:20 +0000 (0:00:00.306)       0:05:17.070 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove old architecture-dependent repository] ****",
                                "Tuesday 13 January 2026  00:29:20 +0000 (0:00:00.409)       0:05:17.479 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather package facts] ****************************",
                                "Tuesday 13 January 2026  00:29:21 +0000 (0:00:00.982)       0:05:18.462 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Check whether packages are installed that should not be installed] ***",
                                "Tuesday 13 January 2026  00:29:24 +0000 (0:00:02.877)       0:05:21.339 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install apt-transport-https package] *************",
                                "Tuesday 13 January 2026  00:29:25 +0000 (0:00:00.566)       0:05:21.906 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository gpg key] **************************",
                                "Tuesday 13 January 2026  00:29:32 +0000 (0:00:07.060)       0:05:28.967 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository] **********************************",
                                "Tuesday 13 January 2026  00:29:33 +0000 (0:00:01.162)       0:05:30.129 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Update package cache] ****************************",
                                "Tuesday 13 January 2026  00:29:41 +0000 (0:00:08.540)       0:05:38.669 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker package version] **********************",
                                "Tuesday 13 January 2026  00:29:45 +0000 (0:00:03.441)       0:05:42.111 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker-cli package version] ******************",
                                "Tuesday 13 January 2026  00:29:46 +0000 (0:00:01.401)       0:05:43.512 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unlock containerd package] ***********************",
                                "Tuesday 13 January 2026  00:29:48 +0000 (0:00:01.520)       0:05:45.033 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install containerd package] **********************",
                                "Tuesday 13 January 2026  00:29:48 +0000 (0:00:00.627)       0:05:45.660 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Lock containerd package] *************************",
                                "Tuesday 13 January 2026  00:29:59 +0000 (0:00:10.733)       0:05:56.394 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker-cli package] **********************",
                                "Tuesday 13 January 2026  00:30:00 +0000 (0:00:00.904)       0:05:57.299 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker package] **************************",
                                "Tuesday 13 January 2026  00:30:11 +0000 (0:00:10.399)       0:06:07.699 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unblock installation of python docker packages] ***",
                                "Tuesday 13 January 2026  00:30:23 +0000 (0:00:12.549)       0:06:20.248 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package] ******************",
                                "Tuesday 13 January 2026  00:30:24 +0000 (0:00:01.334)       0:06:21.583 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package from Debian Sid] ***",
                                "Tuesday 13 January 2026  00:30:25 +0000 (0:00:00.521)       0:06:22.104 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***",
                                "Tuesday 13 January 2026  00:30:29 +0000 (0:00:04.013)       0:06:26.118 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***",
                                "Tuesday 13 January 2026  00:30:29 +0000 (0:00:00.494)       0:06:26.612 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***",
                                "Tuesday 13 January 2026  00:30:30 +0000 (0:00:00.714)       0:06:27.327 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***",
                                "Tuesday 13 January 2026  00:30:31 +0000 (0:00:00.489)       0:06:27.816 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install packages required by docker login] *******",
                                "Tuesday 13 January 2026  00:30:31 +0000 (0:00:00.478)       0:06:28.295 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Ensure that some packages are not installed] *****",
                                "Tuesday 13 January 2026  00:30:32 +0000 (0:00:00.531)       0:06:28.827 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include config tasks] ****************************",
                                "Tuesday 13 January 2026  00:30:34 +0000 (0:00:02.211)       0:06:31.038 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create plugins directory] ************************",
                                "Tuesday 13 January 2026  00:30:35 +0000 (0:00:00.825)       0:06:31.864 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create systemd overlay directory] ****************",
                                "Tuesday 13 January 2026  00:30:36 +0000 (0:00:00.848)       0:06:32.712 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy systemd overlay file] ***********************",
                                "Tuesday 13 January 2026  00:30:36 +0000 (0:00:00.868)       0:06:33.581 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***",
                                "Tuesday 13 January 2026  00:30:38 +0000 (0:00:01.517)       0:06:35.099 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy limits configuration file] ******************",
                                "Tuesday 13 January 2026  00:30:39 +0000 (0:00:01.480)       0:06:36.579 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy daemon.json configuration file] *************",
                                "Tuesday 13 January 2026  00:30:41 +0000 (0:00:01.282)       0:06:37.862 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include service tasks] ***************************",
                                "Tuesday 13 January 2026  00:30:42 +0000 (0:00:01.413)       0:06:39.276 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon] ***************************",
                                "Tuesday 13 January 2026  00:30:43 +0000 (0:00:01.003)       0:06:40.279 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage service] **********************************",
                                "Tuesday 13 January 2026  00:30:45 +0000 (0:00:01.420)       0:06:41.699 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage docker socket service] ********************",
                                "Tuesday 13 January 2026  00:30:46 +0000 (0:00:01.211)       0:06:42.911 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage containerd service] ***********************",
                                "Tuesday 13 January 2026  00:30:47 +0000 (0:00:01.182)       0:06:44.093 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include bootstrap tasks] *************************",
                                "Tuesday 13 January 2026  00:30:48 +0000 (0:00:01.309)       0:06:45.403 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.860)       0:06:46.264 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.039)       0:06:46.304 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.037)       0:06:46.342 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.044)       0:06:46.386 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.038)       0:06:46.424 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.037)       0:06:46.462 ******* ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.044)       0:06:46.507 ******* ",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Tuesday 13 January 2026  00:30:49 +0000 (0:00:00.038)       0:06:46.545 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************",
                                "Tuesday 13 January 2026  00:30:51 +0000 (0:00:01.380)       0:06:47.926 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********",
                                "Tuesday 13 January 2026  00:30:52 +0000 (0:00:01.482)       0:06:49.408 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Restart docker service] ***************",
                                "Tuesday 13 January 2026  00:30:54 +0000 (0:00:01.423)       0:06:50.832 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****",
                                "Tuesday 13 January 2026  00:30:56 +0000 (0:00:02.309)       0:06:53.141 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add user to docker group] ************************",
                                "Tuesday 13 January 2026  00:30:56 +0000 (0:00:00.105)       0:06:53.247 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Log into private registry and force re-authorization] ***",
                                "Tuesday 13 January 2026  00:30:57 +0000 (0:00:01.066)       0:06:54.313 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include facts tasks] *****************************",
                                "Tuesday 13 January 2026  00:30:58 +0000 (0:00:00.506)       0:06:54.819 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create facts directory] **************************",
                                "Tuesday 13 January 2026  00:30:59 +0000 (0:00:01.037)       0:06:55.857 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy docker fact files] **************************",
                                "Tuesday 13 January 2026  00:31:00 +0000 (0:00:00.837)       0:06:56.695 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : This install type is not supported] *******",
                                "Tuesday 13 January 2026  00:31:02 +0000 (0:00:02.566)       0:06:59.261 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Include distribution specific install tasks] ***",
                                "Tuesday 13 January 2026  00:31:03 +0000 (0:00:00.634)       0:06:59.895 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***",
                                "Tuesday 13 January 2026  00:31:03 +0000 (0:00:00.771)       0:07:00.667 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******",
                                "Tuesday 13 January 2026  00:31:04 +0000 (0:00:00.828)       0:07:01.495 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose binary] *************",
                                "Tuesday 13 January 2026  00:31:05 +0000 (0:00:01.033)       0:07:02.528 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Uninstall docker-compose package] *********",
                                "Tuesday 13 January 2026  00:31:06 +0000 (0:00:00.472)       0:07:03.001 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose script] ***************",
                                "Tuesday 13 January 2026  00:31:07 +0000 (0:00:01.625)       0:07:04.626 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****",
                                "Tuesday 13 January 2026  00:31:08 +0000 (0:00:00.481)       0:07:05.108 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********",
                                "Tuesday 13 January 2026  00:31:16 +0000 (0:00:08.415)       0:07:13.523 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Enable osism.target] **********************",
                                "Tuesday 13 January 2026  00:31:18 +0000 (0:00:01.527)       0:07:15.051 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****",
                                "Tuesday 13 January 2026  00:31:20 +0000 (0:00:01.890)       0:07:16.942 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  00:31:21 +0000 (0:00:01.637)       0:07:18.579 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 13 January 2026  00:31:22 +0000 (0:00:00.952)       0:07:19.531 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check minimum and maximum number of servers] *****",
                                "Tuesday 13 January 2026  00:31:23 +0000 (0:00:00.969)       0:07:20.501 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Gather variables for each operating system] ******",
                                "Tuesday 13 January 2026  00:31:24 +0000 (0:00:00.552)       0:07:21.053 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***",
                                "Tuesday 13 January 2026  00:31:24 +0000 (0:00:00.512)       0:07:21.566 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_key_file variable to default value] ***",
                                "Tuesday 13 January 2026  00:31:25 +0000 (0:00:00.530)       0:07:22.096 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Populate service facts] **************************",
                                "Tuesday 13 January 2026  00:31:26 +0000 (0:00:00.712)       0:07:22.808 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage timesyncd service] ************************",
                                "Tuesday 13 January 2026  00:31:31 +0000 (0:00:05.550)       0:07:28.359 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Include distribution specific install tasks] *****",
                                "Tuesday 13 January 2026  00:31:32 +0000 (0:00:00.608)       0:07:28.967 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Install package] *********************************",
                                "Tuesday 13 January 2026  00:31:33 +0000 (0:00:01.134)       0:07:30.102 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage chrony service] ***************************",
                                "Tuesday 13 January 2026  00:31:35 +0000 (0:00:01.982)       0:07:32.084 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check if configuration file exists] **************",
                                "Tuesday 13 January 2026  00:31:36 +0000 (0:00:01.186)       0:07:33.271 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Copy configuration file] *************************",
                                "Tuesday 13 January 2026  00:31:37 +0000 (0:00:00.854)       0:07:34.126 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Include distribution specific install tasks] ******",
                                "Tuesday 13 January 2026  00:31:39 +0000 (0:00:01.937)       0:07:36.063 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Install lldpd package] ****************************",
                                "Tuesday 13 January 2026  00:31:40 +0000 (0:00:00.900)       0:07:36.963 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Manage lldpd service] *****************************",
                                "Tuesday 13 January 2026  00:31:49 +0000 (0:00:09.644)       0:07:46.607 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********",
                                "Tuesday 13 January 2026  00:31:51 +0000 (0:00:01.880)       0:07:48.488 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************",
                                "Tuesday 13 January 2026  00:31:53 +0000 (0:00:01.367)       0:07:49.855 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap role part 2] *********************************************",
                                "",
                                "TASK [Include hardening role] **************************************************",
                                "Tuesday 13 January 2026  00:31:54 +0000 (0:00:01.277)       0:07:51.133 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 3] ********************************************",
                                "",
                                "TASK [osism.services.journald : Copy configuration file] ***********************",
                                "Tuesday 13 January 2026  00:31:55 +0000 (0:00:00.668)       0:07:51.802 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.journald : Manage journald service] ***********************",
                                "Tuesday 13 January 2026  00:31:56 +0000 (0:00:01.433)       0:07:53.235 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include auditd role] *****************************************************",
                                "Tuesday 13 January 2026  00:31:57 +0000 (0:00:01.414)       0:07:54.650 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include smartd role] *****************************************************",
                                "Tuesday 13 January 2026  00:31:58 +0000 (0:00:00.475)       0:07:55.125 ******* ",
                                "\u001b[0;36mincluded: osism.services.smartd for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Include distribution specific install tasks] *****",
                                "Tuesday 13 January 2026  00:31:59 +0000 (0:00:00.918)       0:07:56.043 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Install smartmontools package] *******************",
                                "Tuesday 13 January 2026  00:32:00 +0000 (0:00:00.758)       0:07:56.802 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Create /var/log/smartd directory] ****************",
                                "Tuesday 13 January 2026  00:32:09 +0000 (0:00:09.030)       0:08:05.832 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Copy smartmontools configuration file] ***********",
                                "Tuesday 13 January 2026  00:32:10 +0000 (0:00:01.053)       0:08:06.886 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Manage smartd service] ***************************",
                                "Tuesday 13 January 2026  00:32:11 +0000 (0:00:01.344)       0:08:08.231 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.journald : Restart journald service] ***********",
                                "Tuesday 13 January 2026  00:32:13 +0000 (0:00:01.861)       0:08:10.093 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************",
                                "Tuesday 13 January 2026  00:32:14 +0000 (0:00:01.425)       0:08:11.518 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Set state bootstrap] *****************************************************",
                                "",
                                "TASK [Set osism.bootstrap.status fact] *****************************************",
                                "Tuesday 13 January 2026  00:32:15 +0000 (0:00:01.103)       0:08:12.621 ******* ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  00:32:16 +0000 (0:00:00.782)       0:08:13.404 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Tuesday 13 January 2026  00:32:17 +0000 (0:00:01.089)       0:08:14.494 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set osism.bootstrap.timestamp fact] **************************************",
                                "Tuesday 13 January 2026  00:32:18 +0000 (0:00:01.122)       0:08:15.616 ******* ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  00:32:19 +0000 (0:00:00.795)       0:08:16.412 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Tuesday 13 January 2026  00:32:20 +0000 (0:00:01.081)       0:08:17.494 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:32:21 +0000 (0:00:01.072)       0:08:18.566 ******* ",
                                "=============================================================================== ",
                                "osism.commons.packages : Install required packages --------------------- 77.35s",
                                "osism.commons.packages : Download required packages -------------------- 38.02s",
                                "osism.commons.cleanup : Cleanup installed packages --------------------- 34.30s",
                                "osism.commons.repository : Update package cache ------------------------ 16.14s",
                                "osism.services.docker : Install docker package ------------------------- 12.55s",
                                "osism.commons.systohc : Install util-linux-extra package --------------- 11.70s",
                                "osism.services.docker : Install containerd package --------------------- 10.73s",
                                "osism.services.docker : Install docker-cli package --------------------- 10.40s",
                                "osism.commons.packages : Remove dependencies that are no longer required -- 10.15s",
                                "osism.services.lldpd : Install lldpd package ---------------------------- 9.64s",
                                "osism.services.smartd : Install smartmontools package ------------------- 9.03s",
                                "osism.commons.cleanup : Remove cloudinit package ------------------------ 8.84s",
                                "osism.services.rng : Install rng package -------------------------------- 8.55s",
                                "osism.services.docker : Add repository ---------------------------------- 8.54s",
                                "osism.commons.docker_compose : Install docker-compose-plugin package ---- 8.42s",
                                "osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.17s",
                                "osism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 7.95s",
                                "osism.services.docker : Install apt-transport-https package ------------- 7.06s",
                                "osism.commons.cleanup : Remove dependencies that are no longer required --- 6.62s",
                                "osism.services.chrony : Populate service facts -------------------------- 5.55s",
                                "+ osism apply fail2ban",
                                "\u001b[32m2026-01-13 00:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e47c4f7-4ec2-4007-886b-032cbee6e968 (fail2ban) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:32:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9e47c4f7-4ec2-4007-886b-032cbee6e968 (fail2ban) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role fail2ban] *****************************************************",
                                "",
                                "TASK [osism.services.fail2ban : Include distribution specific install tasks] ***",
                                "Tuesday 13 January 2026  00:32:39 +0000 (0:00:00.275)       0:00:00.275 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Install fail2ban package] **********************",
                                "Tuesday 13 January 2026  00:32:40 +0000 (0:00:01.136)       0:00:01.411 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Copy configuration files] **********************",
                                "Tuesday 13 January 2026  00:32:52 +0000 (0:00:11.264)       0:00:12.676 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Manage fail2ban service] ***********************",
                                "Tuesday 13 January 2026  00:32:53 +0000 (0:00:01.506)       0:00:14.183 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Reload fail2ban configuration] *****************",
                                "Tuesday 13 January 2026  00:32:55 +0000 (0:00:01.581)       0:00:15.764 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:32:56 +0000 (0:00:01.654)       0:00:17.419 ******* ",
                                "=============================================================================== ",
                                "osism.services.fail2ban : Install fail2ban package --------------------- 11.26s",
                                "osism.services.fail2ban : Reload fail2ban configuration ----------------- 1.65s",
                                "osism.services.fail2ban : Manage fail2ban service ----------------------- 1.58s",
                                "osism.services.fail2ban : Copy configuration files ---------------------- 1.51s",
                                "osism.services.fail2ban : Include distribution specific install tasks --- 1.14s",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ osism apply network",
                                "\u001b[32m2026-01-13 00:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 08e96ea9-e9e0-489d-99af-6eaeeff5e332 (network) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:33:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 08e96ea9-e9e0-489d-99af-6eaeeff5e332 (network) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role network] ******************************************************",
                                "",
                                "TASK [osism.commons.network : Gather variables for each operating system] ******",
                                "Tuesday 13 January 2026  00:33:13 +0000 (0:00:00.255)       0:00:00.255 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include type specific tasks] *********************",
                                "Tuesday 13 January 2026  00:33:14 +0000 (0:00:00.568)       0:00:00.823 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install required packages] ***********************",
                                "Tuesday 13 January 2026  00:33:15 +0000 (0:00:01.005)       0:00:01.828 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove ifupdown package] *************************",
                                "Tuesday 13 January 2026  00:33:17 +0000 (0:00:02.040)       0:00:03.869 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create required directories] *********************",
                                "Tuesday 13 January 2026  00:33:19 +0000 (0:00:01.735)       0:00:05.605 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Prepare netplan configuration template] **********",
                                "Tuesday 13 January 2026  00:33:20 +0000 (0:00:00.957)       0:00:06.562 ******* ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy netplan configuration] **********************",
                                "Tuesday 13 January 2026  00:33:23 +0000 (0:00:03.244)       0:00:09.806 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove netplan configuration template] ***********",
                                "Tuesday 13 January 2026  00:33:24 +0000 (0:00:01.552)       0:00:11.359 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Check if path for interface file exists] *********",
                                "Tuesday 13 January 2026  00:33:26 +0000 (0:00:01.614)       0:00:12.974 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy interfaces file] ****************************",
                                "Tuesday 13 January 2026  00:33:27 +0000 (0:00:01.114)       0:00:14.088 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install package networkd-dispatcher] *************",
                                "Tuesday 13 January 2026  00:33:28 +0000 (0:00:00.648)       0:00:14.737 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy dispatcher scripts] *************************",
                                "Tuesday 13 January 2026  00:33:30 +0000 (0:00:02.203)       0:00:16.940 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Manage service networkd-dispatcher] **************",
                                "Tuesday 13 January 2026  00:33:31 +0000 (0:00:00.902)       0:00:17.842 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include cleanup tasks] ***************************",
                                "Tuesday 13 January 2026  00:33:33 +0000 (0:00:01.627)       0:00:19.470 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Tuesday 13 January 2026  00:33:34 +0000 (0:00:01.193)       0:00:20.664 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Set network_configured_files fact] ***************",
                                "Tuesday 13 January 2026  00:33:35 +0000 (0:00:01.104)       0:00:21.769 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Tuesday 13 January 2026  00:33:36 +0000 (0:00:00.626)       0:00:22.395 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include dummy interfaces] ************************",
                                "Tuesday 13 January 2026  00:33:37 +0000 (0:00:01.192)       0:00:23.588 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include vxlan interfaces] ************************",
                                "Tuesday 13 January 2026  00:33:37 +0000 (0:00:00.586)       0:00:24.175 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-0, testbed-node-2, testbed-manager, testbed-node-3, testbed-node-1, testbed-node-5, testbed-node-4\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd netdev files] ************",
                                "Tuesday 13 January 2026  00:33:42 +0000 (0:00:04.307)       0:00:28.482 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd network files] ***********",
                                "Tuesday 13 January 2026  00:33:47 +0000 (0:00:05.341)       0:00:33.824 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include networkd cleanup tasks] ******************",
                                "Tuesday 13 January 2026  00:33:52 +0000 (0:00:05.033)       0:00:38.858 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Tuesday 13 January 2026  00:33:53 +0000 (0:00:01.071)       0:00:39.930 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Tuesday 13 January 2026  00:33:54 +0000 (0:00:01.010)       0:00:40.941 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include network extra init] **********************",
                                "Tuesday 13 January 2026  00:33:55 +0000 (0:00:00.783)       0:00:41.724 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init script] ****************",
                                "Tuesday 13 January 2026  00:33:56 +0000 (0:00:01.204)       0:00:42.928 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init systemd service] *******",
                                "Tuesday 13 January 2026  00:33:57 +0000 (0:00:00.632)       0:00:43.561 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Enable and start network-extra-init service] *****",
                                "Tuesday 13 January 2026  00:33:57 +0000 (0:00:00.740)       0:00:44.302 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Disable and stop network-extra-init service] *****",
                                "Tuesday 13 January 2026  00:33:58 +0000 (0:00:00.585)       0:00:44.887 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init systemd service] *******",
                                "Tuesday 13 January 2026  00:34:00 +0000 (0:00:01.753)       0:00:46.641 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init script] ****************",
                                "Tuesday 13 January 2026  00:34:01 +0000 (0:00:00.964)       0:00:47.606 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************",
                                "Tuesday 13 January 2026  00:34:03 +0000 (0:00:02.221)       0:00:49.827 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********",
                                "Tuesday 13 January 2026  00:34:04 +0000 (0:00:00.808)       0:00:50.635 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:34:04 +0000 (0:00:00.532)       0:00:51.168 ******* ",
                                "=============================================================================== ",
                                "osism.commons.network : Create systemd networkd netdev files ------------ 5.34s",
                                "osism.commons.network : Create systemd networkd network files ----------- 5.03s",
                                "osism.commons.network : Include vxlan interfaces ------------------------ 4.31s",
                                "osism.commons.network : Prepare netplan configuration template ---------- 3.24s",
                                "osism.commons.network : Remove network-extra-init script ---------------- 2.22s",
                                "osism.commons.network : Install package networkd-dispatcher ------------- 2.20s",
                                "osism.commons.network : Install required packages ----------------------- 2.04s",
                                "osism.commons.network : Disable and stop network-extra-init service ----- 1.75s",
                                "osism.commons.network : Remove ifupdown package ------------------------- 1.74s",
                                "osism.commons.network : Manage service networkd-dispatcher -------------- 1.63s",
                                "osism.commons.network : Remove netplan configuration template ----------- 1.61s",
                                "osism.commons.network : Copy netplan configuration ---------------------- 1.55s",
                                "osism.commons.network : Include network extra init ---------------------- 1.20s",
                                "osism.commons.network : Include cleanup tasks --------------------------- 1.19s",
                                "osism.commons.network : Remove unused configuration files --------------- 1.19s",
                                "osism.commons.network : Check if path for interface file exists --------- 1.11s",
                                "osism.commons.network : List existing configuration files --------------- 1.10s",
                                "osism.commons.network : Include networkd cleanup tasks ------------------ 1.07s",
                                "osism.commons.network : List existing configuration files --------------- 1.01s",
                                "osism.commons.network : Include type specific tasks --------------------- 1.01s",
                                "+ osism apply wireguard",
                                "\u001b[32m2026-01-13 00:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8a352e9e-84ed-40db-bcd1-dca8ac701d87 (wireguard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8a352e9e-84ed-40db-bcd1-dca8ac701d87 (wireguard) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role wireguard] ****************************************************",
                                "",
                                "TASK [osism.services.wireguard : Install iptables package] *********************",
                                "Tuesday 13 January 2026  00:34:21 +0000 (0:00:00.163)       0:00:00.163 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Install wireguard package] ********************",
                                "Tuesday 13 January 2026  00:34:22 +0000 (0:00:01.303)       0:00:01.467 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create public and private key - server] *******",
                                "Tuesday 13 January 2026  00:34:29 +0000 (0:00:06.612)       0:00:08.080 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create preshared key] *************************",
                                "Tuesday 13 January 2026  00:34:29 +0000 (0:00:00.561)       0:00:08.641 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get preshared key] ****************************",
                                "Tuesday 13 January 2026  00:34:30 +0000 (0:00:00.428)       0:00:09.070 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get public key - server] **********************",
                                "Tuesday 13 January 2026  00:34:31 +0000 (0:00:00.684)       0:00:09.754 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get private key - server] *********************",
                                "Tuesday 13 January 2026  00:34:31 +0000 (0:00:00.397)       0:00:10.151 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy wg0.conf configuration file] *************",
                                "Tuesday 13 January 2026  00:34:31 +0000 (0:00:00.405)       0:00:10.557 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy client configuration files] **************",
                                "Tuesday 13 January 2026  00:34:33 +0000 (0:00:01.162)       0:00:11.719 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********",
                                "Tuesday 13 January 2026  00:34:33 +0000 (0:00:00.954)       0:00:12.674 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************",
                                "Tuesday 13 January 2026  00:34:35 +0000 (0:00:01.621)       0:00:14.296 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:34:36 +0000 (0:00:00.899)       0:00:15.195 ******* ",
                                "=============================================================================== ",
                                "osism.services.wireguard : Install wireguard package -------------------- 6.61s",
                                "osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.62s",
                                "osism.services.wireguard : Install iptables package --------------------- 1.30s",
                                "osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.16s",
                                "osism.services.wireguard : Copy client configuration files -------------- 0.95s",
                                "osism.services.wireguard : Restart wg0 service -------------------------- 0.90s",
                                "osism.services.wireguard : Get preshared key ---------------------------- 0.68s",
                                "osism.services.wireguard : Create public and private key - server ------- 0.56s",
                                "osism.services.wireguard : Create preshared key ------------------------- 0.43s",
                                "osism.services.wireguard : Get private key - server --------------------- 0.41s",
                                "osism.services.wireguard : Get public key - server ---------------------- 0.40s",
                                "+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0",
                                "100    15  100    15    0     0    197      0 --:--:-- --:--:-- --:--:--   200",
                                "+ osism apply --environment custom workarounds",
                                "\u001b[32m2026-01-13 00:34:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m",
                                "\u001b[32m2026-01-13 00:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8323183f-4a7a-455e-bd47-9ad302c4b40a (workarounds) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8323183f-4a7a-455e-bd47-9ad302c4b40a (workarounds) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on virtualization_role] ********************************",
                                "Tuesday 13 January 2026  00:34:53 +0000 (0:00:00.125)       0:00:00.125 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on the manager node] *************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Tuesday 13 January 2026  00:34:54 +0000 (0:00:00.774)       0:00:00.899 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on all other nodes] **************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Tuesday 13 January 2026  00:34:56 +0000 (0:00:02.139)       0:00:03.039 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Add custom CA certificates to non-manager nodes] *************************",
                                "",
                                "TASK [Copy custom CA certificates] *********************************************",
                                "Tuesday 13 January 2026  00:34:58 +0000 (0:00:01.996)       0:00:05.036 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "",
                                "TASK [Run update-ca-certificates] **********************************************",
                                "Tuesday 13 January 2026  00:34:59 +0000 (0:00:01.513)       0:00:06.549 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Run update-ca-trust] *****************************************************",
                                "Tuesday 13 January 2026  00:35:03 +0000 (0:00:04.052)       0:00:10.602 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Add a workaround service] ************************************************",
                                "",
                                "TASK [Copy workarounds.sh scripts] *********************************************",
                                "Tuesday 13 January 2026  00:35:04 +0000 (0:00:00.632)       0:00:11.234 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Copy workarounds systemd unit file] **************************************",
                                "Tuesday 13 January 2026  00:35:05 +0000 (0:00:01.459)       0:00:12.694 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Reload systemd daemon] ***************************************************",
                                "Tuesday 13 January 2026  00:35:07 +0000 (0:00:01.475)       0:00:14.169 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable workarounds.service (Debian)] *************************************",
                                "Tuesday 13 January 2026  00:35:08 +0000 (0:00:01.490)       0:00:15.659 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable and start workarounds.service (RedHat)] ***************************",
                                "Tuesday 13 January 2026  00:35:10 +0000 (0:00:01.708)       0:00:17.368 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************",
                                "",
                                "TASK [Install python3-docker] **************************************************",
                                "Tuesday 13 January 2026  00:35:11 +0000 (0:00:00.613)       0:00:17.982 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:35:13 +0000 (0:00:02.735)       0:00:20.717 ******* ",
                                "=============================================================================== ",
                                "Run update-ca-certificates ---------------------------------------------- 4.05s",
                                "Install python3-docker -------------------------------------------------- 2.74s",
                                "Apply netplan configuration --------------------------------------------- 2.14s",
                                "Apply netplan configuration --------------------------------------------- 2.00s",
                                "Enable workarounds.service (Debian) ------------------------------------- 1.71s",
                                "Copy custom CA certificates --------------------------------------------- 1.51s",
                                "Reload systemd daemon --------------------------------------------------- 1.49s",
                                "Copy workarounds systemd unit file -------------------------------------- 1.48s",
                                "Copy workarounds.sh scripts --------------------------------------------- 1.46s",
                                "Group hosts based on virtualization_role -------------------------------- 0.77s",
                                "Run update-ca-trust ----------------------------------------------------- 0.63s",
                                "Enable and start workarounds.service (RedHat) --------------------------- 0.61s",
                                "+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-01-13 00:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dd3467b7-2d05-49ce-b4ed-d7ed65f1d10f (reboot) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dd3467b7-2d05-49ce-b4ed-d7ed65f1d10f (reboot) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:30 +0000 (0:00:00.150)       0:00:00.150 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:30 +0000 (0:00:00.076)       0:00:00.227 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:31 +0000 (0:00:00.918)       0:00:01.145 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:31 +0000 (0:00:00.093)       0:00:01.238 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:31 +0000 (0:00:00.084)       0:00:01.323 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:32 +0000 (0:00:00.660)       0:00:01.983 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:32 +0000 (0:00:00.096)       0:00:02.079 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:32 +0000 (0:00:00.142)       0:00:02.221 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:33 +0000 (0:00:00.645)       0:00:02.867 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:33 +0000 (0:00:00.090)       0:00:02.958 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:33 +0000 (0:00:00.078)       0:00:03.037 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:34 +0000 (0:00:00.653)       0:00:03.690 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:34 +0000 (0:00:00.109)       0:00:03.799 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:34 +0000 (0:00:00.086)       0:00:03.886 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:35 +0000 (0:00:00.736)       0:00:04.623 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 13 January 2026  00:35:35 +0000 (0:00:00.117)       0:00:04.740 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 13 January 2026  00:35:35 +0000 (0:00:00.091)       0:00:04.832 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 13 January 2026  00:35:36 +0000 (0:00:00.696)       0:00:05.528 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:35:36 +0000 (0:00:00.034)       0:00:05.563 ******* ",
                                "=============================================================================== ",
                                "Reboot system - do not wait for the reboot to complete ------------------ 4.31s",
                                "Exit playbook, if user did not mean to reboot systems ------------------- 0.56s",
                                "Reboot system - wait for the reboot to complete ------------------------- 0.54s",
                                "+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-01-13 00:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 71b89266-deb6-462c-a198-3aa3a9843757 (wait-for-connection) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 71b89266-deb6-462c-a198-3aa3a9843757 (wait-for-connection) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wait until remote systems are reachable] *********************************",
                                "",
                                "TASK [Wait until remote system is reachable] ***********************************",
                                "Tuesday 13 January 2026  00:35:52 +0000 (0:00:00.169)       0:00:00.169 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:36:04 +0000 (0:00:11.477)       0:00:11.646 ******* ",
                                "=============================================================================== ",
                                "Wait until remote system is reachable ---------------------------------- 11.48s",
                                "+ osism apply hddtemp",
                                "\u001b[32m2026-01-13 00:36:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ca78f17-8193-46a9-8856-5f474ebd664b (hddtemp) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:36:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7ca78f17-8193-46a9-8856-5f474ebd664b (hddtemp) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role hddtemp] ******************************************************",
                                "",
                                "TASK [osism.services.hddtemp : Gather variables for each operating system] *****",
                                "Tuesday 13 January 2026  00:36:21 +0000 (0:00:00.252)       0:00:00.252 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific install tasks] ****",
                                "Tuesday 13 January 2026  00:36:22 +0000 (0:00:00.684)       0:00:00.936 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Remove hddtemp package] *************************",
                                "Tuesday 13 January 2026  00:36:23 +0000 (0:00:01.155)       0:00:02.092 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************",
                                "Tuesday 13 January 2026  00:36:25 +0000 (0:00:02.214)       0:00:04.307 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Check if drivetemp module is available] *********",
                                "Tuesday 13 January 2026  00:36:26 +0000 (0:00:01.057)       0:00:05.364 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************",
                                "Tuesday 13 January 2026  00:36:27 +0000 (0:00:01.087)       0:00:06.451 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Install lm-sensors] *****************************",
                                "Tuesday 13 January 2026  00:36:28 +0000 (0:00:00.758)       0:00:07.210 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific service tasks] ****",
                                "Tuesday 13 January 2026  00:36:41 +0000 (0:00:13.623)       0:00:20.834 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Manage lm-sensors service] **********************",
                                "Tuesday 13 January 2026  00:36:43 +0000 (0:00:01.165)       0:00:22.000 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:36:44 +0000 (0:00:01.737)       0:00:23.738 ******* ",
                                "=============================================================================== ",
                                "osism.services.hddtemp : Install lm-sensors ---------------------------- 13.62s",
                                "osism.services.hddtemp : Remove hddtemp package ------------------------- 2.21s",
                                "osism.services.hddtemp : Manage lm-sensors service ---------------------- 1.74s",
                                "osism.services.hddtemp : Include distribution specific service tasks ---- 1.17s",
                                "osism.services.hddtemp : Include distribution specific install tasks ---- 1.16s",
                                "osism.services.hddtemp : Check if drivetemp module is available --------- 1.09s",
                                "osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.06s",
                                "osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.76s",
                                "osism.services.hddtemp : Gather variables for each operating system ----- 0.68s",
                                "++ semver latest 7.1.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sudo systemctl restart manager.service",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "ARA in ceph-ansible already disabled.",
                                "ARA in kolla-ansible already disabled.",
                                "ARA in osism-ansible already disabled.",
                                "ARA in osism-kubernetes already disabled.",
                                "+ osism apply gather-facts",
                                "\u001b[32m2026-01-13 00:38:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ce44a44-dd4f-4cb9-89b4-2d880292c235 (gather-facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 6ce44a44-dd4f-4cb9-89b4-2d880292c235 (gather-facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:38:20 +0000 (0:00:00.186)       0:00:00.186 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 13 January 2026  00:38:28 +0000 (0:00:08.414)       0:00:08.600 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:38:29 +0000 (0:00:00.419)       0:00:09.019 ******* ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 8.41s",
                                "Gather facts for all hosts ---------------------------------------------- 0.42s",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry",
                                "+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images",
                                "+ [[ false == \\t\\r\\u\\e ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001a-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:38:30.431665Z",
                            "start": "2026-01-13T00:14:08.345277Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001a",
                        "name": "Deploy manager + bootstrap nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "not _nutshell | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T00:38:31.048104Z",
                            "start": "2026-01-13T00:38:30.537439Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001b",
                        "name": "Deploy services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "/opt/configuration/scripts/deploy-in-a-nutshell.sh"
                            ],
                            "delta": "0:33:48.557797",
                            "end": "2026-01-13 01:12:19.981856",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 /opt/configuration/scripts/deploy-in-a-nutshell.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001c-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 00:38:31.424059",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n\n# PULL IMAGES\n\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# PULL IMAGES'\n+ echo\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-01-13 00:38:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-01-13 00:38:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6e14501f-fcd9-40f9-abe9-92730c7570b3 (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:38:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6e14501f-fcd9-40f9-abe9-92730c7570b3 is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-01-13 00:38:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m\n\u001b[32m2026-01-13 00:38:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2c21fb30-2dda-4d49-a0de-66a5504a8069 (wipe-partitions) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:38:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2c21fb30-2dda-4d49-a0de-66a5504a8069 (wipe-partitions) has been started and output is visible here.\u001b[0m\n\nPLAY [Wipe partitions] *********************************************************\n\nTASK [Find all logical devices owned by UID 167] *******************************\nTuesday 13 January 2026  00:39:00 +0000 (0:00:00.134)       0:00:00.134 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Remove all rook related logical devices] *********************************\nTuesday 13 January 2026  00:39:01 +0000 (0:00:00.600)       0:00:00.735 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Find all logical devices with prefix ceph] *******************************\nTuesday 13 January 2026  00:39:01 +0000 (0:00:00.325)       0:00:01.060 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Remove all ceph related logical devices] *********************************\nTuesday 13 January 2026  00:39:02 +0000 (0:00:00.568)       0:00:01.629 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check device availability] ***********************************************\nTuesday 13 January 2026  00:39:02 +0000 (0:00:00.238)       0:00:01.867 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Wipe partitions with wipefs] *********************************************\nTuesday 13 January 2026  00:39:03 +0000 (0:00:01.160)       0:00:03.028 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Overwrite first 32M with zeros] ******************************************\nTuesday 13 January 2026  00:39:05 +0000 (0:00:01.475)       0:00:04.503 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\nTASK [Reload udev rules] *******************************************************\nTuesday 13 January 2026  00:39:09 +0000 (0:00:04.019)       0:00:08.523 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Request device events from the kernel] ***********************************\nTuesday 13 January 2026  00:39:09 +0000 (0:00:00.596)       0:00:09.120 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:39:10 +0000 (0:00:00.602)       0:00:09.722 ******* \n=============================================================================== \nOverwrite first 32M with zeros ------------------------------------------ 4.02s\nWipe partitions with wipefs --------------------------------------------- 1.48s\nCheck device availability ----------------------------------------------- 1.16s\nRequest device events from the kernel ----------------------------------- 0.60s\nFind all logical devices owned by UID 167 ------------------------------- 0.60s\nReload udev rules ------------------------------------------------------- 0.60s\nFind all logical devices with prefix ceph ------------------------------- 0.57s\nRemove all rook related logical devices --------------------------------- 0.33s\nRemove all ceph related logical devices --------------------------------- 0.24s\n\u001b[32m2026-01-13 00:39:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f01a7713-5f5f-42c7-b15b-1d9d3ceefd85 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:39:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f01a7713-5f5f-42c7-b15b-1d9d3ceefd85 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 13 January 2026  00:39:27 +0000 (0:00:00.254)       0:00:00.254 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 13 January 2026  00:39:28 +0000 (0:00:00.912)       0:00:01.166 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:39:29 +0000 (0:00:01.042)       0:00:02.209 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 13 January 2026  00:39:33 +0000 (0:00:04.791)       0:00:07.000 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:39:34 +0000 (0:00:00.495)       0:00:07.496 ******* \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.79s\nosism.commons.facts : Copy fact files ----------------------------------- 1.04s\nosism.commons.facts : Create custom facts directory --------------------- 0.91s\nGather facts for all hosts ---------------------------------------------- 0.50s\n\u001b[32m2026-01-13 00:39:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1aaf4672-483a-4030-a225-6812d7e6d7f1 (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:39:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1aaf4672-483a-4030-a225-6812d7e6d7f1 (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:39:41 +0000 (0:00:00.305)       0:00:00.305 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:39:41 +0000 (0:00:00.229)       0:00:00.535 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:41 +0000 (0:00:00.208)       0:00:00.743 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:42 +0000 (0:00:00.461)       0:00:01.205 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:42 +0000 (0:00:00.187)       0:00:01.393 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:42 +0000 (0:00:00.168)       0:00:01.561 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:42 +0000 (0:00:00.179)       0:00:01.741 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:42 +0000 (0:00:00.176)       0:00:01.917 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:43 +0000 (0:00:00.185)       0:00:02.102 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:43 +0000 (0:00:00.185)       0:00:02.288 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:43 +0000 (0:00:00.193)       0:00:02.482 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:43 +0000 (0:00:00.183)       0:00:02.666 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:44 +0000 (0:00:00.425)       0:00:03.091 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:44 +0000 (0:00:00.576)       0:00:03.668 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:45 +0000 (0:00:00.565)       0:00:04.234 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:46 +0000 (0:00:00.770)       0:00:05.005 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:46 +0000 (0:00:00.332)       0:00:05.337 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:46 +0000 (0:00:00.350)       0:00:05.688 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:46 +0000 (0:00:00.189)       0:00:05.877 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:47 +0000 (0:00:00.188)       0:00:06.066 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:47 +0000 (0:00:00.181)       0:00:06.247 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:47 +0000 (0:00:00.182)       0:00:06.429 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:47 +0000 (0:00:00.180)       0:00:06.610 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:47 +0000 (0:00:00.188)       0:00:06.798 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:48 +0000 (0:00:00.200)       0:00:06.998 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:48 +0000 (0:00:00.189)       0:00:07.188 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:49 +0000 (0:00:00.919)       0:00:08.107 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:49 +0000 (0:00:00.180)       0:00:08.288 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:49 +0000 (0:00:00.185)       0:00:08.473 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:49 +0000 (0:00:00.190)       0:00:08.663 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 13 January 2026  00:39:49 +0000 (0:00:00.192)       0:00:08.855 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.163)       0:00:09.018 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.121)       0:00:09.139 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.122)       0:00:09.262 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.112)       0:00:09.375 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.125)       0:00:09.500 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.153)       0:00:09.654 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 13 January 2026  00:39:50 +0000 (0:00:00.140)       0:00:09.795 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.308)       0:00:10.103 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.136)       0:00:10.240 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.129)       0:00:10.370 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.145)       0:00:10.515 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.133)       0:00:10.648 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.115)       0:00:10.764 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 13 January 2026  00:39:51 +0000 (0:00:00.136)       0:00:10.900 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 13 January 2026  00:39:52 +0000 (0:00:00.141)       0:00:11.041 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 13 January 2026  00:39:52 +0000 (0:00:00.134)       0:00:11.176 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 13 January 2026  00:39:52 +0000 (0:00:00.131)       0:00:11.307 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 13 January 2026  00:39:52 +0000 (0:00:00.124)       0:00:11.431 ******* \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-03961d85-1922-5669-8251-0ccc6cca9fac\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 13 January 2026  00:39:52 +0000 (0:00:00.375)       0:00:11.807 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:39:54 +0000 (0:00:01.731)       0:00:13.538 ******* \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:39:54 +0000 (0:00:00.250)       0:00:13.789 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:55 +0000 (0:00:00.222)       0:00:14.011 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:55 +0000 (0:00:00.345)       0:00:14.357 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:55 +0000 (0:00:00.172)       0:00:14.529 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:55 +0000 (0:00:00.182)       0:00:14.712 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:55 +0000 (0:00:00.186)       0:00:14.898 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:56 +0000 (0:00:00.184)       0:00:15.083 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:56 +0000 (0:00:00.508)       0:00:15.591 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:56 +0000 (0:00:00.180)       0:00:15.772 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:56 +0000 (0:00:00.184)       0:00:15.956 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:57 +0000 (0:00:00.188)       0:00:16.144 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:57 +0000 (0:00:00.392)       0:00:16.537 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:57 +0000 (0:00:00.390)       0:00:16.928 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:58 +0000 (0:00:00.404)       0:00:17.333 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:39:58 +0000 (0:00:00.395)       0:00:17.729 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:59 +0000 (0:00:00.322)       0:00:18.051 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:39:59 +0000 (0:00:00.354)       0:00:18.406 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:00 +0000 (0:00:00.669)       0:00:19.076 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:00 +0000 (0:00:00.191)       0:00:19.268 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:00 +0000 (0:00:00.218)       0:00:19.486 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:00 +0000 (0:00:00.201)       0:00:19.687 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:00 +0000 (0:00:00.195)       0:00:19.883 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:01 +0000 (0:00:00.208)       0:00:20.092 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:01 +0000 (0:00:00.190)       0:00:20.282 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:01 +0000 (0:00:00.187)       0:00:20.470 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:02 +0000 (0:00:00.686)       0:00:21.157 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:02 +0000 (0:00:00.156)       0:00:21.314 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:02 +0000 (0:00:00.166)       0:00:21.480 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:02 +0000 (0:00:00.177)       0:00:21.657 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.481)       0:00:22.139 ******* \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.110)       0:00:22.249 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.116)       0:00:22.365 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.103)       0:00:22.469 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.105)       0:00:22.575 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.106)       0:00:22.681 ******* \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.133)       0:00:22.815 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 13 January 2026  00:40:03 +0000 (0:00:00.116)       0:00:22.932 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.121)       0:00:23.054 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.114)       0:00:23.168 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.105)       0:00:23.274 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.102)       0:00:23.376 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.237)       0:00:23.614 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.136)       0:00:23.750 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 13 January 2026  00:40:04 +0000 (0:00:00.099)       0:00:23.849 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 13 January 2026  00:40:05 +0000 (0:00:00.142)       0:00:23.993 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 13 January 2026  00:40:05 +0000 (0:00:00.099)       0:00:24.092 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 13 January 2026  00:40:05 +0000 (0:00:00.109)       0:00:24.202 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 13 January 2026  00:40:05 +0000 (0:00:00.102)       0:00:24.304 ******* \n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 13 January 2026  00:40:05 +0000 (0:00:00.180)       0:00:24.484 ******* \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:40:06 +0000 (0:00:00.995)       0:00:25.480 ******* \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:40:07 +0000 (0:00:00.482)       0:00:25.963 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:07 +0000 (0:00:00.194)       0:00:26.157 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:07 +0000 (0:00:00.307)       0:00:26.464 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:07 +0000 (0:00:00.175)       0:00:26.640 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:07 +0000 (0:00:00.176)       0:00:26.817 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.180)       0:00:26.997 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.170)       0:00:27.168 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.187)       0:00:27.355 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.166)       0:00:27.522 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.153)       0:00:27.676 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:08 +0000 (0:00:00.169)       0:00:27.846 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:09 +0000 (0:00:00.537)       0:00:28.383 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:09 +0000 (0:00:00.365)       0:00:28.749 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:10 +0000 (0:00:00.400)       0:00:29.149 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:40:10 +0000 (0:00:00.469)       0:00:29.618 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:10 +0000 (0:00:00.318)       0:00:29.937 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:11 +0000 (0:00:00.383)       0:00:30.320 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:11 +0000 (0:00:00.181)       0:00:30.502 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:11 +0000 (0:00:00.203)       0:00:30.706 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:11 +0000 (0:00:00.180)       0:00:30.886 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:12 +0000 (0:00:00.181)       0:00:31.068 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:12 +0000 (0:00:00.160)       0:00:31.228 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:12 +0000 (0:00:00.393)       0:00:31.621 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:12 +0000 (0:00:00.172)       0:00:31.794 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:12 +0000 (0:00:00.165)       0:00:31.959 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:13 +0000 (0:00:00.554)       0:00:32.514 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:13 +0000 (0:00:00.166)       0:00:32.681 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:13 +0000 (0:00:00.167)       0:00:32.848 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.173)       0:00:33.021 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.230)       0:00:33.252 ******* \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.191)       0:00:33.443 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.139)       0:00:33.582 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.131)       0:00:33.714 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 13 January 2026  00:40:14 +0000 (0:00:00.236)       0:00:33.950 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.109)       0:00:34.059 ******* \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.131)       0:00:34.190 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.121)       0:00:34.312 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.126)       0:00:34.438 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.136)       0:00:34.575 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.132)       0:00:34.707 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.109)       0:00:34.816 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 13 January 2026  00:40:15 +0000 (0:00:00.101)       0:00:34.918 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.131)       0:00:35.049 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.112)       0:00:35.162 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.120)       0:00:35.282 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.232)       0:00:35.514 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.100)       0:00:35.614 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.100)       0:00:35.715 ******* \n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-e91d200a-cf56-55df-b2f8-08f15361112f\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 13 January 2026  00:40:16 +0000 (0:00:00.169)       0:00:35.885 ******* \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:40:17 +0000 (0:00:00.852)       0:00:36.738 ******* \n=============================================================================== \nWrite configuration file ------------------------------------------------ 3.58s\nAdd known links to the list of available block devices ------------------ 1.11s\nAdd known partitions to the list of available block devices ------------- 1.09s\nGet extra vars for Ceph configuration ----------------------------------- 0.96s\nAdd known partitions to the list of available block devices ------------- 0.92s\nAdd known links to the list of available block devices ------------------ 0.77s\nPrint configuration data ------------------------------------------------ 0.73s\nAdd known partitions to the list of available block devices ------------- 0.69s\nAdd known partitions to the list of available block devices ------------- 0.67s\nGet initial list of available block devices ----------------------------- 0.63s\nAdd known links to the list of available block devices ------------------ 0.58s\nAdd known links to the list of available block devices ------------------ 0.57s\nGenerate lvm_volumes structure (block + wal) ---------------------------- 0.56s\nAdd known partitions to the list of available block devices ------------- 0.55s\nAdd known links to the list of available block devices ------------------ 0.54s\nAdd known links to the list of available block devices ------------------ 0.51s\nAdd known partitions to the list of available block devices ------------- 0.48s\nSet DB devices config data ---------------------------------------------- 0.47s\nAdd known links to the list of available block devices ------------------ 0.47s\nPrint WAL devices ------------------------------------------------------- 0.47s\n\u001b[32m2026-01-13 00:40:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5b6e9c71-ef81-48c9-b389-576a03ff85e9 (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-01-13 00:40:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-01-13 00:40:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-01-13 00:40:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-01-13 00:40:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-01-13 00:41:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n[master ea9e2af] 2026-01-13-00-41\n 1 file changed, 30 insertions(+), 9 deletions(-)\n\u001b[32m2026-01-13 00:41:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a91a25af-a18a-484c-9019-a75474dc5746 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:41:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a91a25af-a18a-484c-9019-a75474dc5746 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:41:12 +0000 (0:00:00.283)       0:00:00.283 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:41:12 +0000 (0:00:00.195)       0:00:00.479 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:12 +0000 (0:00:00.181)       0:00:00.661 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:13 +0000 (0:00:00.412)       0:00:01.073 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:13 +0000 (0:00:00.171)       0:00:01.245 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:13 +0000 (0:00:00.170)       0:00:01.415 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:13 +0000 (0:00:00.184)       0:00:01.600 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:14 +0000 (0:00:00.193)       0:00:01.794 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:14 +0000 (0:00:00.159)       0:00:01.953 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:14 +0000 (0:00:00.129)       0:00:02.083 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:14 +0000 (0:00:00.151)       0:00:02.234 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:14 +0000 (0:00:00.162)       0:00:02.396 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:15 +0000 (0:00:00.380)       0:00:02.777 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:15 +0000 (0:00:00.570)       0:00:03.347 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:16 +0000 (0:00:00.512)       0:00:03.860 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:16 +0000 (0:00:00.692)       0:00:04.553 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:17 +0000 (0:00:00.299)       0:00:04.852 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:17 +0000 (0:00:00.368)       0:00:05.221 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:17 +0000 (0:00:00.183)       0:00:05.405 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:17 +0000 (0:00:00.179)       0:00:05.584 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:18 +0000 (0:00:00.168)       0:00:05.753 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:18 +0000 (0:00:00.184)       0:00:05.937 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:18 +0000 (0:00:00.211)       0:00:06.148 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:18 +0000 (0:00:00.208)       0:00:06.357 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:18 +0000 (0:00:00.231)       0:00:06.589 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:19 +0000 (0:00:00.232)       0:00:06.821 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:20 +0000 (0:00:01.173)       0:00:07.995 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:20 +0000 (0:00:00.191)       0:00:08.186 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:20 +0000 (0:00:00.175)       0:00:08.362 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:20 +0000 (0:00:00.198)       0:00:08.560 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 13 January 2026  00:41:21 +0000 (0:00:00.202)       0:00:08.763 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 13 January 2026  00:41:21 +0000 (0:00:00.130)       0:00:08.893 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 13 January 2026  00:41:21 +0000 (0:00:00.175)       0:00:09.068 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 13 January 2026  00:41:23 +0000 (0:00:01.926)       0:00:10.995 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 13 January 2026  00:41:23 +0000 (0:00:00.145)       0:00:11.140 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 13 January 2026  00:41:24 +0000 (0:00:01.450)       0:00:12.591 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.162)       0:00:12.753 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.127)       0:00:12.880 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.364)       0:00:13.244 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.126)       0:00:13.371 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.135)       0:00:13.507 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 13 January 2026  00:41:25 +0000 (0:00:00.132)       0:00:13.639 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.162)       0:00:13.802 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.121)       0:00:13.923 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.134)       0:00:14.058 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.145)       0:00:14.203 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.154)       0:00:14.358 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.118)       0:00:14.476 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 13 January 2026  00:41:26 +0000 (0:00:00.125)       0:00:14.601 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 13 January 2026  00:41:27 +0000 (0:00:00.113)       0:00:14.715 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 13 January 2026  00:41:27 +0000 (0:00:00.296)       0:00:15.011 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 13 January 2026  00:41:27 +0000 (0:00:00.157)       0:00:15.169 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 13 January 2026  00:41:27 +0000 (0:00:00.159)       0:00:15.328 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 13 January 2026  00:41:28 +0000 (0:00:00.601)       0:00:15.930 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 13 January 2026  00:41:28 +0000 (0:00:00.509)       0:00:16.440 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.525)       0:00:16.965 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.127)       0:00:17.093 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.126)       0:00:17.219 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.122)       0:00:17.341 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.127)       0:00:17.469 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 13 January 2026  00:41:29 +0000 (0:00:00.133)       0:00:17.602 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 13 January 2026  00:41:30 +0000 (0:00:00.157)       0:00:17.760 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 13 January 2026  00:41:30 +0000 (0:00:00.462)       0:00:18.223 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 13 January 2026  00:41:30 +0000 (0:00:00.199)       0:00:18.422 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 13 January 2026  00:41:30 +0000 (0:00:00.161)       0:00:18.583 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.138)       0:00:18.722 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.152)       0:00:18.874 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.185)       0:00:19.060 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.142)       0:00:19.202 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.142)       0:00:19.345 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.157)       0:00:19.502 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 13 January 2026  00:41:31 +0000 (0:00:00.144)       0:00:19.646 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:41:32 +0000 (0:00:00.152)       0:00:19.799 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 13 January 2026  00:41:32 +0000 (0:00:00.158)       0:00:19.957 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 13 January 2026  00:41:32 +0000 (0:00:00.486)       0:00:20.443 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 13 January 2026  00:41:32 +0000 (0:00:00.172)       0:00:20.616 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.177)       0:00:20.794 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.173)       0:00:20.968 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.161)       0:00:21.129 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.161)       0:00:21.291 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.171)       0:00:21.462 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 13 January 2026  00:41:33 +0000 (0:00:00.150)       0:00:21.613 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 13 January 2026  00:41:34 +0000 (0:00:00.498)       0:00:22.111 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 13 January 2026  00:41:34 +0000 (0:00:00.544)       0:00:22.656 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 13 January 2026  00:41:35 +0000 (0:00:00.149)       0:00:22.806 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'vg_name': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'vg_name': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 13 January 2026  00:41:35 +0000 (0:00:00.166)       0:00:22.972 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 13 January 2026  00:41:35 +0000 (0:00:00.377)       0:00:23.349 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 13 January 2026  00:41:35 +0000 (0:00:00.167)       0:00:23.516 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 13 January 2026  00:41:35 +0000 (0:00:00.173)       0:00:23.690 ******* \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-03961d85-1922-5669-8251-0ccc6cca9fac\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:41:36 +0000 (0:00:00.317)       0:00:24.007 ******* \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:41:36 +0000 (0:00:00.235)       0:00:24.242 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:36 +0000 (0:00:00.238)       0:00:24.480 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:37 +0000 (0:00:00.497)       0:00:24.978 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:37 +0000 (0:00:00.201)       0:00:25.179 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:37 +0000 (0:00:00.215)       0:00:25.395 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:38 +0000 (0:00:00.874)       0:00:26.270 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:38 +0000 (0:00:00.258)       0:00:26.528 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:39 +0000 (0:00:00.233)       0:00:26.762 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:39 +0000 (0:00:00.216)       0:00:26.978 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:39 +0000 (0:00:00.230)       0:00:27.209 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:39 +0000 (0:00:00.178)       0:00:27.387 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:40 +0000 (0:00:00.420)       0:00:27.808 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:40 +0000 (0:00:00.408)       0:00:28.217 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:40 +0000 (0:00:00.399)       0:00:28.616 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:41:41 +0000 (0:00:00.614)       0:00:29.231 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:42 +0000 (0:00:00.537)       0:00:29.769 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:42 +0000 (0:00:00.838)       0:00:30.607 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:43 +0000 (0:00:00.202)       0:00:30.810 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:43 +0000 (0:00:00.202)       0:00:31.013 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:43 +0000 (0:00:00.198)       0:00:31.211 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:43 +0000 (0:00:00.196)       0:00:31.408 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:43 +0000 (0:00:00.194)       0:00:31.603 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:44 +0000 (0:00:00.200)       0:00:31.804 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:44 +0000 (0:00:00.214)       0:00:32.019 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:44 +0000 (0:00:00.190)       0:00:32.209 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:45 +0000 (0:00:00.831)       0:00:33.040 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:45 +0000 (0:00:00.178)       0:00:33.219 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:46 +0000 (0:00:00.619)       0:00:33.838 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:41:46 +0000 (0:00:00.223)       0:00:34.062 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 13 January 2026  00:41:46 +0000 (0:00:00.206)       0:00:34.269 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 13 January 2026  00:41:46 +0000 (0:00:00.149)       0:00:34.418 ******* \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 13 January 2026  00:41:46 +0000 (0:00:00.199)       0:00:34.617 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 13 January 2026  00:41:48 +0000 (0:00:01.893)       0:00:36.511 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 13 January 2026  00:41:48 +0000 (0:00:00.149)       0:00:36.661 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 13 January 2026  00:41:50 +0000 (0:00:01.382)       0:00:38.043 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 13 January 2026  00:41:50 +0000 (0:00:00.157)       0:00:38.201 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 13 January 2026  00:41:50 +0000 (0:00:00.152)       0:00:38.354 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 13 January 2026  00:41:50 +0000 (0:00:00.144)       0:00:38.499 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 13 January 2026  00:41:50 +0000 (0:00:00.146)       0:00:38.645 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 13 January 2026  00:41:51 +0000 (0:00:00.368)       0:00:39.014 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 13 January 2026  00:41:51 +0000 (0:00:00.140)       0:00:39.155 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 13 January 2026  00:41:51 +0000 (0:00:00.161)       0:00:39.317 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 13 January 2026  00:41:51 +0000 (0:00:00.141)       0:00:39.458 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 13 January 2026  00:41:51 +0000 (0:00:00.166)       0:00:39.624 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.157)       0:00:39.782 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.155)       0:00:39.937 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.126)       0:00:40.064 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.142)       0:00:40.206 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.131)       0:00:40.338 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.133)       0:00:40.471 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 13 January 2026  00:41:52 +0000 (0:00:00.135)       0:00:40.607 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 13 January 2026  00:41:53 +0000 (0:00:00.348)       0:00:40.955 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 13 January 2026  00:41:53 +0000 (0:00:00.545)       0:00:41.500 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 13 January 2026  00:41:54 +0000 (0:00:00.556)       0:00:42.057 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 13 January 2026  00:41:54 +0000 (0:00:00.533)       0:00:42.590 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.160)       0:00:42.751 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.106)       0:00:42.858 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.103)       0:00:42.961 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.151)       0:00:43.112 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.132)       0:00:43.245 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.127)       0:00:43.373 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.127)       0:00:43.501 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 13 January 2026  00:41:55 +0000 (0:00:00.124)       0:00:43.625 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.335)       0:00:43.960 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.139)       0:00:44.100 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.128)       0:00:44.229 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.133)       0:00:44.362 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.133)       0:00:44.496 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:41:56 +0000 (0:00:00.143)       0:00:44.639 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.147)       0:00:44.787 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.148)       0:00:44.935 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.134)       0:00:45.069 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.148)       0:00:45.218 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.164)       0:00:45.383 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 13 January 2026  00:41:57 +0000 (0:00:00.145)       0:00:45.529 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 13 January 2026  00:41:58 +0000 (0:00:00.347)       0:00:45.876 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 13 January 2026  00:41:58 +0000 (0:00:00.158)       0:00:46.035 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 13 January 2026  00:41:58 +0000 (0:00:00.175)       0:00:46.211 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 13 January 2026  00:41:58 +0000 (0:00:00.154)       0:00:46.366 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 13 January 2026  00:41:58 +0000 (0:00:00.179)       0:00:46.545 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 13 January 2026  00:41:59 +0000 (0:00:00.169)       0:00:46.715 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 13 January 2026  00:41:59 +0000 (0:00:00.501)       0:00:47.216 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.515)       0:00:47.732 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.158)       0:00:47.890 ******* \n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'vg_name': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'vg_name': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.195)       0:00:48.086 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.173)       0:00:48.259 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.176)       0:00:48.435 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 13 January 2026  00:42:00 +0000 (0:00:00.176)       0:00:48.612 ******* \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  00:42:01 +0000 (0:00:00.545)       0:00:49.157 ******* \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 13 January 2026  00:42:01 +0000 (0:00:00.259)       0:00:49.416 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:01 +0000 (0:00:00.250)       0:00:49.666 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:02 +0000 (0:00:00.429)       0:00:50.096 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:02 +0000 (0:00:00.236)       0:00:50.332 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:02 +0000 (0:00:00.176)       0:00:50.509 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:03 +0000 (0:00:00.192)       0:00:50.701 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:03 +0000 (0:00:00.209)       0:00:50.911 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:03 +0000 (0:00:00.592)       0:00:51.503 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:04 +0000 (0:00:00.200)       0:00:51.704 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:04 +0000 (0:00:00.222)       0:00:51.927 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:04 +0000 (0:00:00.193)       0:00:52.120 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:04 +0000 (0:00:00.393)       0:00:52.514 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:05 +0000 (0:00:00.406)       0:00:52.920 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:05 +0000 (0:00:00.446)       0:00:53.366 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 13 January 2026  00:42:06 +0000 (0:00:00.445)       0:00:53.811 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:06 +0000 (0:00:00.354)       0:00:54.166 ******* \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:06 +0000 (0:00:00.403)       0:00:54.569 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:07 +0000 (0:00:00.196)       0:00:54.765 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:07 +0000 (0:00:00.776)       0:00:55.542 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:08 +0000 (0:00:00.195)       0:00:55.738 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:08 +0000 (0:00:00.195)       0:00:55.933 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:08 +0000 (0:00:00.183)       0:00:56.117 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:08 +0000 (0:00:00.191)       0:00:56.308 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:08 +0000 (0:00:00.204)       0:00:56.512 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:09 +0000 (0:00:00.208)       0:00:56.721 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:09 +0000 (0:00:00.732)       0:00:57.453 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:09 +0000 (0:00:00.215)       0:00:57.669 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:10 +0000 (0:00:00.204)       0:00:57.874 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 13 January 2026  00:42:10 +0000 (0:00:00.186)       0:00:58.060 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 13 January 2026  00:42:10 +0000 (0:00:00.220)       0:00:58.280 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 13 January 2026  00:42:10 +0000 (0:00:00.291)       0:00:58.572 ******* \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 13 January 2026  00:42:11 +0000 (0:00:00.180)       0:00:58.753 ******* \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 13 January 2026  00:42:12 +0000 (0:00:01.810)       0:01:00.564 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 13 January 2026  00:42:13 +0000 (0:00:00.149)       0:01:00.713 ******* \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 13 January 2026  00:42:14 +0000 (0:00:01.301)       0:01:02.014 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 13 January 2026  00:42:14 +0000 (0:00:00.156)       0:01:02.171 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 13 January 2026  00:42:14 +0000 (0:00:00.125)       0:01:02.296 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 13 January 2026  00:42:14 +0000 (0:00:00.159)       0:01:02.456 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 13 January 2026  00:42:14 +0000 (0:00:00.130)       0:01:02.586 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 13 January 2026  00:42:15 +0000 (0:00:00.151)       0:01:02.738 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 13 January 2026  00:42:15 +0000 (0:00:00.142)       0:01:02.880 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 13 January 2026  00:42:15 +0000 (0:00:00.147)       0:01:03.028 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 13 January 2026  00:42:15 +0000 (0:00:00.363)       0:01:03.391 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 13 January 2026  00:42:15 +0000 (0:00:00.168)       0:01:03.560 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.166)       0:01:03.726 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.163)       0:01:03.890 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.141)       0:01:04.032 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.132)       0:01:04.164 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.157)       0:01:04.321 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.133)       0:01:04.454 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 13 January 2026  00:42:16 +0000 (0:00:00.155)       0:01:04.610 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 13 January 2026  00:42:17 +0000 (0:00:00.140)       0:01:04.750 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 13 January 2026  00:42:17 +0000 (0:00:00.512)       0:01:05.262 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 13 January 2026  00:42:18 +0000 (0:00:00.531)       0:01:05.794 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 13 January 2026  00:42:18 +0000 (0:00:00.735)       0:01:06.530 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 13 January 2026  00:42:18 +0000 (0:00:00.154)       0:01:06.684 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.118)       0:01:06.802 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.094)       0:01:06.897 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.123)       0:01:07.020 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.116)       0:01:07.136 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.129)       0:01:07.266 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.155)       0:01:07.422 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.131)       0:01:07.553 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 13 January 2026  00:42:19 +0000 (0:00:00.131)       0:01:07.684 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.124)       0:01:07.809 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.126)       0:01:07.936 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.338)       0:01:08.275 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.132)       0:01:08.407 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.132)       0:01:08.540 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 13 January 2026  00:42:20 +0000 (0:00:00.119)       0:01:08.660 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.136)       0:01:08.797 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.122)       0:01:08.919 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.119)       0:01:09.039 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.146)       0:01:09.186 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.146)       0:01:09.332 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.148)       0:01:09.480 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 13 January 2026  00:42:21 +0000 (0:00:00.143)       0:01:09.623 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 13 January 2026  00:42:22 +0000 (0:00:00.147)       0:01:09.771 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 13 January 2026  00:42:22 +0000 (0:00:00.358)       0:01:10.129 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 13 January 2026  00:42:22 +0000 (0:00:00.164)       0:01:10.294 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 13 January 2026  00:42:22 +0000 (0:00:00.145)       0:01:10.439 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 13 January 2026  00:42:23 +0000 (0:00:00.525)       0:01:10.965 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 13 January 2026  00:42:23 +0000 (0:00:00.521)       0:01:11.487 ******* \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 13 January 2026  00:42:23 +0000 (0:00:00.133)       0:01:11.621 ******* \n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'vg_name': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'vg_name': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 13 January 2026  00:42:24 +0000 (0:00:00.165)       0:01:11.787 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 13 January 2026  00:42:24 +0000 (0:00:00.148)       0:01:11.936 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 13 January 2026  00:42:24 +0000 (0:00:00.151)       0:01:12.087 ******* \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 13 January 2026  00:42:24 +0000 (0:00:00.165)       0:01:12.252 ******* \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-e91d200a-cf56-55df-b2f8-08f15361112f\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:42:24 +0000 (0:00:00.140)       0:01:12.393 ******* \n=============================================================================== \nCreate block VGs -------------------------------------------------------- 5.63s\nCreate block LVs -------------------------------------------------------- 4.13s\nGather DB+WAL VGs with total and available size in bytes ---------------- 1.79s\nGather DB VGs with total and available size in bytes -------------------- 1.66s\nAdd known partitions to the list of available block devices ------------- 1.61s\nGather WAL VGs with total and available size in bytes ------------------- 1.60s\nGet list of Ceph PVs with associated VGs -------------------------------- 1.58s\nGet list of Ceph LVs with associated VGs -------------------------------- 1.53s\nAdd known links to the list of available block devices ------------------ 1.34s\nAdd known partitions to the list of available block devices ------------- 1.17s\nPrint LVM report data --------------------------------------------------- 1.00s\nAdd known links to the list of available block devices ------------------ 0.87s\nAdd known partitions to the list of available block devices ------------- 0.83s\nCreate DB LVs for ceph_db_devices --------------------------------------- 0.80s\nAdd known partitions to the list of available block devices ------------- 0.78s\nPrint size needed for LVs on ceph_db_devices ---------------------------- 0.75s\nAdd known partitions to the list of available block devices ------------- 0.73s\nFail if block LV defined in lvm_volumes is missing ---------------------- 0.70s\nAdd known links to the list of available block devices ------------------ 0.69s\nGet extra vars for Ceph configuration ----------------------------------- 0.69s\n\u001b[32m2026-01-13 00:42:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f227694-2bc5-4186-89a0-6a2c85f9a271 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:42:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2f227694-2bc5-4186-89a0-6a2c85f9a271 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 13 January 2026  00:42:41 +0000 (0:00:00.272)       0:00:00.272 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 13 January 2026  00:42:42 +0000 (0:00:01.049)       0:00:01.322 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  00:42:43 +0000 (0:00:01.226)       0:00:02.548 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 13 January 2026  00:42:48 +0000 (0:00:04.955)       0:00:07.503 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:42:49 +0000 (0:00:00.482)       0:00:07.985 ******* \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.96s\nosism.commons.facts : Copy fact files ----------------------------------- 1.23s\nosism.commons.facts : Create custom facts directory --------------------- 1.05s\nGather facts for all hosts ---------------------------------------------- 0.48s\n\u001b[32m2026-01-13 00:43:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be4c30f5-fe4a-4450-a3bd-83c6374cf20b (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 00:43:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task be4c30f5-fe4a-4450-a3bd-83c6374cf20b (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nTuesday 13 January 2026  00:43:06 +0000 (0:00:00.265)       0:00:00.265 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nTuesday 13 January 2026  00:43:07 +0000 (0:00:00.253)       0:00:00.519 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nTuesday 13 January 2026  00:43:08 +0000 (0:00:01.301)       0:00:01.821 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nTuesday 13 January 2026  00:43:18 +0000 (0:00:10.520)       0:00:12.341 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nTuesday 13 January 2026  00:43:20 +0000 (0:00:01.017)       0:00:13.359 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nTuesday 13 January 2026  00:43:20 +0000 (0:00:00.973)       0:00:14.333 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nTuesday 13 January 2026  00:43:22 +0000 (0:00:01.200)       0:00:15.533 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nTuesday 13 January 2026  00:43:22 +0000 (0:00:00.163)       0:00:15.697 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nTuesday 13 January 2026  00:43:22 +0000 (0:00:00.163)       0:00:15.861 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nTuesday 13 January 2026  00:43:23 +0000 (0:00:00.982)       0:00:16.844 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nTuesday 13 January 2026  00:43:26 +0000 (0:00:03.331)       0:00:20.176 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.frr : Restart frr service] *********************\nTuesday 13 January 2026  00:43:28 +0000 (0:00:01.694)       0:00:21.870 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:43:30 +0000 (0:00:01.494)       0:00:23.365 ******* \n=============================================================================== \nosism.services.frr : Install frr package ------------------------------- 10.52s\nosism.services.frr : Set sysctl parameters ------------------------------ 3.33s\nosism.services.frr : Manage frr service --------------------------------- 1.69s\nosism.services.frr : Restart frr service -------------------------------- 1.49s\nosism.services.frr : Pin frr package version ---------------------------- 1.30s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 1.20s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.02s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.98s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.97s\nosism.services.frr : Include distribution specific install tasks -------- 0.25s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 0.16s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 0.16s\n\n--> DEPLOY IN A NUTSHELL -- START -- Tue Jan 13 00:43:30 UTC 2026\n\n\u001b[32m2026-01-13 00:43:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m\n\u001b[32m2026-01-13 00:43:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m\n\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m\n\u001b[32m2026-01-13 00:43:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m\n\u001b[32m2026-01-13 00:43:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m\n\u001b[32m2026-01-13 00:43:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\nPLAY [Apply role geerlingguy.dotfiles] *****************************************\n\nTASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****\nTuesday 13 January 2026  00:43:54 +0000 (0:00:00.446)       0:00:00.446 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********\nTuesday 13 January 2026  00:43:58 +0000 (0:00:03.847)       0:00:04.293 ******* \n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***\nTuesday 13 January 2026  00:44:00 +0000 (0:00:02.188)       0:00:06.481 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.814918', 'end': '2026-01-13 00:43:58.820429', 'delta': '0:00:00.005511', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.803859', 'end': '2026-01-13 00:43:58.813355', 'delta': '0:00:00.009496', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.817015', 'end': '2026-01-13 00:43:58.826967', 'delta': '0:00:00.009952', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.124810', 'end': '2026-01-13 00:43:59.132657', 'delta': '0:00:00.007847', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.630323', 'end': '2026-01-13 00:43:59.637507', 'delta': '0:00:00.007184', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.371314', 'end': '2026-01-13 00:43:59.375075', 'delta': '0:00:00.003761', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:44:00.067080', 'end': '2026-01-13 00:44:00.075232', 'delta': '0:00:00.008152', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****\nTuesday 13 January 2026  00:44:02 +0000 (0:00:01.866)       0:00:08.348 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************\nTuesday 13 January 2026  00:44:04 +0000 (0:00:02.001)       0:00:10.349 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:44:07 +0000 (0:00:03.360)       0:00:13.710 ******* \n=============================================================================== \ngeerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 3.85s\ngeerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 3.36s\ngeerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.19s\ngeerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 2.00s\ngeerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 1.87s\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role homer] ********************************************************\n\nTASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***\nTuesday 13 January 2026  00:43:55 +0000 (0:00:01.241)       0:00:01.241 ******* \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.services.homer : Create traefik external network] ******************\nTuesday 13 January 2026  00:43:55 +0000 (0:00:00.241)       0:00:01.482 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Create required directories] **********************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:01.068)       0:00:02.551 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m\n\nTASK [osism.services.homer : Copy config.yml configuration file] ***************\nTuesday 13 January 2026  00:43:58 +0000 (0:00:01.720)       0:00:04.272 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Copy docker-compose.yml file] *********************\nTuesday 13 January 2026  00:44:01 +0000 (0:00:02.720)       0:00:06.992 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Manage homer service] *****************************\nTuesday 13 January 2026  00:44:03 +0000 (0:00:01.731)       0:00:08.723 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.homer : Restart homer service] *****************\nTuesday 13 January 2026  00:44:31 +0000 (0:00:28.445)       0:00:37.169 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:44:35 +0000 (0:00:04.266)       0:00:41.435 ******* \n=============================================================================== \nosism.services.homer : Manage homer service ---------------------------- 28.45s\nosism.services.homer : Restart homer service ---------------------------- 4.27s\nosism.services.homer : Copy config.yml configuration file --------------- 2.72s\nosism.services.homer : Copy docker-compose.yml file --------------------- 1.73s\nosism.services.homer : Create required directories ---------------------- 1.72s\nosism.services.homer : Create traefik external network ------------------ 1.07s\nosism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.24s\n\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:00.581)       0:00:00.581 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:00.359)       0:00:00.940 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nTuesday 13 January 2026  00:43:59 +0000 (0:00:02.825)       0:00:03.766 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nTuesday 13 January 2026  00:44:02 +0000 (0:00:03.162)       0:00:06.929 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nTuesday 13 January 2026  00:44:36 +0000 (0:00:34.142)       0:00:41.071 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nTuesday 13 January 2026  00:44:38 +0000 (0:00:01.743)       0:00:42.815 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nTuesday 13 January 2026  00:44:39 +0000 (0:00:01.001)       0:00:43.816 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nTuesday 13 January 2026  00:44:42 +0000 (0:00:02.612)       0:00:46.429 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nTuesday 13 January 2026  00:44:44 +0000 (0:00:02.059)       0:00:48.488 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nTuesday 13 January 2026  00:44:44 +0000 (0:00:00.648)       0:00:49.137 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.413)       0:00:49.550 ******* \n=============================================================================== \nosism.services.openstackclient : Manage openstackclient service -------- 34.14s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 3.16s\nosism.services.openstackclient : Create required directories ------------ 2.83s\nosism.services.openstackclient : Restart openstackclient service -------- 2.61s\nosism.services.openstackclient : Ensure that all containers are up ------ 2.06s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 1.74s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 1.00s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.65s\nosism.services.openstackclient : Copy bash completion script ------------ 0.41s\nosism.services.openstackclient : Include tasks -------------------------- 0.36s\n\n\nPLAY [Apply role phpmyadmin] ***************************************************\n\nTASK [osism.services.phpmyadmin : Create traefik external network] *************\nTuesday 13 January 2026  00:44:12 +0000 (0:00:00.354)       0:00:00.354 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Create required directories] *****************\nTuesday 13 January 2026  00:44:14 +0000 (0:00:01.132)       0:00:01.487 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m\n\nTASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************\nTuesday 13 January 2026  00:44:14 +0000 (0:00:00.602)       0:00:02.089 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************\nTuesday 13 January 2026  00:44:16 +0000 (0:00:01.619)       0:00:03.708 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******\nTuesday 13 January 2026  00:45:09 +0000 (0:00:53.611)       0:00:57.320 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:45:15 +0000 (0:00:05.575)       0:01:02.895 ******* \n=============================================================================== \nosism.services.phpmyadmin : Manage phpmyadmin service ------------------ 53.61s\nosism.services.phpmyadmin : Restart phpmyadmin service ------------------ 5.58s\nosism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.62s\nosism.services.phpmyadmin : Create traefik external network ------------- 1.13s\nosism.services.phpmyadmin : Create required directories ----------------- 0.60s\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:43:55 +0000 (0:00:00.305)       0:00:00.305 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m\n\nPLAY [Apply role netdata] ******************************************************\n\nTASK [osism.services.netdata : Include distribution specific install tasks] ****\nTuesday 13 January 2026  00:43:56 +0000 (0:00:01.322)       0:00:01.628 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Remove old architecture-dependent repository] ***\nTuesday 13 January 2026  00:43:59 +0000 (0:00:03.084)       0:00:04.713 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Install apt-transport-https package] ************\nTuesday 13 January 2026  00:44:02 +0000 (0:00:02.505)       0:00:07.218 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add repository gpg key] *************************\nTuesday 13 January 2026  00:44:05 +0000 (0:00:03.685)       0:00:10.904 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Add repository] *********************************\nTuesday 13 January 2026  00:44:08 +0000 (0:00:02.058)       0:00:12.963 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Install package netdata] ************************\nTuesday 13 January 2026  00:44:20 +0000 (0:00:12.305)       0:00:25.269 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Include config tasks] ***************************\nTuesday 13 January 2026  00:45:04 +0000 (0:00:43.787)       0:01:09.057 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Copy configuration files] ***********************\nTuesday 13 January 2026  00:45:05 +0000 (0:00:01.388)       0:01:10.445 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m\n\nTASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***\nTuesday 13 January 2026  00:45:10 +0000 (0:00:04.818)       0:01:15.263 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Opt out from anonymous statistics] **************\nTuesday 13 January 2026  00:45:11 +0000 (0:00:00.968)       0:01:16.232 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add netdata user to docker group] ***************\nTuesday 13 January 2026  00:45:12 +0000 (0:00:01.259)       0:01:17.491 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Manage service netdata] *************************\nTuesday 13 January 2026  00:45:13 +0000 (0:00:01.343)       0:01:18.835 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Include host type specific tasks] ***************\nTuesday 13 January 2026  00:45:16 +0000 (0:00:02.361)       0:01:21.197 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********\nTuesday 13 January 2026  00:45:18 +0000 (0:00:01.823)       0:01:23.020 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.netdata : Restart service netdata] *************\nTuesday 13 January 2026  00:45:19 +0000 (0:00:01.897)       0:01:24.918 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:45:23 +0000 (0:00:03.431)       0:01:28.349 ******* \n=============================================================================== \nosism.services.netdata : Install package netdata ----------------------- 43.79s\nosism.services.netdata : Add repository -------------------------------- 12.31s\nosism.services.netdata : Copy configuration files ----------------------- 4.82s\nosism.services.netdata : Install apt-transport-https package ------------ 3.69s\nosism.services.netdata : Restart service netdata ------------------------ 3.43s\nosism.services.netdata : Include distribution specific install tasks ---- 3.08s\nosism.services.netdata : Remove old architecture-dependent repository --- 2.51s\nosism.services.netdata : Manage service netdata ------------------------- 2.36s\nosism.services.netdata : Add repository gpg key ------------------------- 2.06s\nosism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.90s\nosism.services.netdata : Include host type specific tasks --------------- 1.82s\nosism.services.netdata : Include config tasks --------------------------- 1.39s\nosism.services.netdata : Add netdata user to docker group --------------- 1.34s\nGroup hosts based on enabled services ----------------------------------- 1.32s\nosism.services.netdata : Opt out from anonymous statistics -------------- 1.26s\nosism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 0.97s\n\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nTuesday 13 January 2026  00:43:47 +0000 (0:00:00.249)       0:00:00.249 ******* \n\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  00:43:49 +0000 (0:00:01.242)       0:00:01.491 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nTuesday 13 January 2026  00:43:53 +0000 (0:00:03.835)       0:00:05.326 ******* \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nTuesday 13 January 2026  00:43:54 +0000 (0:00:01.122)       0:00:06.449 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:43:58 +0000 (0:00:04.707)       0:00:11.157 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nTuesday 13 January 2026  00:44:00 +0000 (0:00:01.538)       0:00:12.696 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nTuesday 13 January 2026  00:44:03 +0000 (0:00:03.058)       0:00:15.754 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nTuesday 13 January 2026  00:44:04 +0000 (0:00:00.896)       0:00:16.650 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nTuesday 13 January 2026  00:44:05 +0000 (0:00:01.115)       0:00:17.765 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nTuesday 13 January 2026  00:44:11 +0000 (0:00:06.384)       0:00:24.149 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nTuesday 13 January 2026  00:44:14 +0000 (0:00:02.343)       0:00:26.493 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nTuesday 13 January 2026  00:44:15 +0000 (0:00:01.110)       0:00:27.603 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nTuesday 13 January 2026  00:44:16 +0000 (0:00:00.987)       0:00:28.591 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nTuesday 13 January 2026  00:44:17 +0000 (0:00:00.872)       0:00:29.464 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nTuesday 13 January 2026  00:44:20 +0000 (0:00:03.315)       0:00:32.779 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nTuesday 13 January 2026  00:44:23 +0000 (0:00:03.119)       0:00:35.898 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nTuesday 13 January 2026  00:44:27 +0000 (0:00:03.634)       0:00:39.533 ******* \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nTuesday 13 January 2026  00:44:31 +0000 (0:00:04.745)       0:00:44.278 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nTuesday 13 January 2026  00:44:34 +0000 (0:00:02.433)       0:00:46.711 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [common : Check common containers] ****************************************\nTuesday 13 January 2026  00:44:37 +0000 (0:00:02.608)       0:00:49.320 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Creating log volume] ********************************************\nTuesday 13 January 2026  00:44:40 +0000 (0:00:03.331)       0:00:52.651 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Link kolla_logs volume to /var/log/kolla] ***********************\nTuesday 13 January 2026  00:44:42 +0000 (0:00:02.177)       0:00:54.829 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:44 +0000 (0:00:01.932)       0:00:56.761 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:44 +0000 (0:00:00.077)       0:00:56.838 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:44 +0000 (0:00:00.080)       0:00:56.919 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.413)       0:00:57.332 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.128)       0:00:57.461 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.121)       0:00:57.583 ******* \n\nTASK [common : Flush handlers] *************************************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.078)       0:00:57.661 ******* \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nTuesday 13 January 2026  00:44:45 +0000 (0:00:00.097)       0:00:57.759 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nTuesday 13 January 2026  00:45:16 +0000 (0:00:30.903)       0:01:28.662 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nTuesday 13 January 2026  00:45:50 +0000 (0:00:34.589)       0:02:03.252 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nTuesday 13 January 2026  00:45:53 +0000 (0:00:02.078)       0:02:05.331 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:46:02 +0000 (0:00:09.755)       0:02:15.086 ******* \n=============================================================================== \ncommon : Restart kolla-toolbox container ------------------------------- 34.59s\ncommon : Restart fluentd container ------------------------------------- 30.90s\ncommon : Restart cron container ----------------------------------------- 9.76s\ncommon : Copying over config.json files for services -------------------- 6.38s\ncommon : Ensuring config directories have correct owner and permission --- 4.75s\nservice-cert-copy : common | Copying over extra CA certificates --------- 4.71s\ncommon : Ensuring config directories exist ------------------------------ 3.84s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 3.63s\ncommon : Check common containers ---------------------------------------- 3.33s\ncommon : Copying over fluentd.conf -------------------------------------- 3.32s\ncommon : Copying over cron logrotate config file ------------------------ 3.12s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 3.06s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.61s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.43s\ncommon : Find custom fluentd input config files ------------------------- 2.34s\ncommon : Creating log volume -------------------------------------------- 2.18s\ncommon : Initializing toolbox container using normal user --------------- 2.08s\ncommon : Link kolla_logs volume to /var/log/kolla ----------------------- 1.93s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 1.54s\ncommon : include_tasks -------------------------------------------------- 1.24s\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.227)       0:00:00.227 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.270)       0:00:00.497 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.416)       0:00:00.914 ******* \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nTuesday 13 January 2026  00:46:10 +0000 (0:00:00.495)       0:00:01.409 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nTuesday 13 January 2026  00:46:10 +0000 (0:00:00.634)       0:00:02.043 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\nTASK [memcached : Check memcached container] ***********************************\nTuesday 13 January 2026  00:46:12 +0000 (0:00:01.907)       0:00:03.951 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nTuesday 13 January 2026  00:46:14 +0000 (0:00:01.570)       0:00:05.521 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:46:22 +0000 (0:00:08.193)       0:00:13.715 ******* \n=============================================================================== \nmemcached : Restart memcached container --------------------------------- 8.19s\nmemcached : Copying over config.json files for services ----------------- 1.91s\nmemcached : Check memcached container ----------------------------------- 1.57s\nmemcached : Ensuring config directories exist --------------------------- 0.63s\nmemcached : include_tasks ----------------------------------------------- 0.50s\nGroup hosts based on enabled services ----------------------------------- 0.42s\nGroup hosts based on Kolla action --------------------------------------- 0.27s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.216)       0:00:00.217 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.236)       0:00:00.453 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.396)       0:00:00.850 ******* \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.463)       0:00:01.314 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nTuesday 13 January 2026  00:46:10 +0000 (0:00:01.204)       0:00:02.519 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nTuesday 13 January 2026  00:46:13 +0000 (0:00:02.515)       0:00:05.034 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Check redis containers] ******************************************\nTuesday 13 January 2026  00:46:15 +0000 (0:00:02.372)       0:00:07.406 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nTuesday 13 January 2026  00:46:17 +0000 (0:00:02.044)       0:00:09.451 ******* \n\nTASK [redis : Flush handlers] **************************************************\nTuesday 13 January 2026  00:46:17 +0000 (0:00:00.109)       0:00:09.561 ******* \n\nTASK [redis : Flush handlers] **************************************************\nTuesday 13 January 2026  00:46:18 +0000 (0:00:00.104)       0:00:09.666 ******* \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nTuesday 13 January 2026  00:46:18 +0000 (0:00:00.090)       0:00:09.756 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nTuesday 13 January 2026  00:46:27 +0000 (0:00:09.003)       0:00:18.759 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:46:37 +0000 (0:00:09.987)       0:00:28.747 ******* \n=============================================================================== \nredis : Restart redis-sentinel container -------------------------------- 9.99s\nredis : Restart redis container ----------------------------------------- 9.00s\nredis : Copying over default config.json files -------------------------- 2.52s\nredis : Copying over redis config files --------------------------------- 2.37s\nredis : Check redis containers ------------------------------------------ 2.05s\nredis : Ensuring config directories exist ------------------------------- 1.20s\nredis : include_tasks --------------------------------------------------- 0.46s\nGroup hosts based on enabled services ----------------------------------- 0.40s\nredis : Flush handlers -------------------------------------------------- 0.30s\nGroup hosts based on Kolla action --------------------------------------- 0.24s\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.202)       0:00:00.202 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.690)       0:00:00.893 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.567)       0:00:01.461 ******* \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 13 January 2026  00:46:10 +0000 (0:00:01.174)       0:00:02.635 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 13 January 2026  00:46:12 +0000 (0:00:01.393)       0:00:04.029 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 13 January 2026  00:46:13 +0000 (0:00:01.466)       0:00:05.496 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nTuesday 13 January 2026  00:46:14 +0000 (0:00:01.101)       0:00:06.597 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nTuesday 13 January 2026  00:46:15 +0000 (0:00:00.770)       0:00:07.368 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nTuesday 13 January 2026  00:46:17 +0000 (0:00:01.851)       0:00:09.220 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nTuesday 13 January 2026  00:46:20 +0000 (0:00:02.935)       0:00:12.156 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Check openvswitch containers] ******************************\nTuesday 13 January 2026  00:46:21 +0000 (0:00:00.990)       0:00:13.146 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:24 +0000 (0:00:02.915)       0:00:16.062 ******* \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:24 +0000 (0:00:00.662)       0:00:16.724 ******* \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.266)       0:00:16.991 ******* \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.324)       0:00:17.315 ******* \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.356)       0:00:17.671 ******* \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 13 January 2026  00:46:26 +0000 (0:00:00.277)       0:00:17.948 ******* \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nTuesday 13 January 2026  00:46:26 +0000 (0:00:00.278)       0:00:18.227 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nTuesday 13 January 2026  00:46:37 +0000 (0:00:10.747)       0:00:28.974 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nTuesday 13 January 2026  00:46:38 +0000 (0:00:01.130)       0:00:30.104 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nTuesday 13 January 2026  00:46:47 +0000 (0:00:09.252)       0:00:39.357 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nTuesday 13 January 2026  00:46:55 +0000 (0:00:07.869)       0:00:47.226 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nTuesday 13 January 2026  00:46:58 +0000 (0:00:03.046)       0:00:50.273 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nTuesday 13 January 2026  00:47:02 +0000 (0:00:03.928)       0:00:54.201 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:47:10 +0000 (0:00:07.944)       0:01:02.146 ******* \n=============================================================================== \nopenvswitch : Restart openvswitch-vswitchd container ------------------- 17.20s\nopenvswitch : Restart openvswitch-db-server container ------------------ 10.75s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 7.87s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 3.93s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 3.05s\nopenvswitch : Copying over config.json files for services --------------- 2.94s\nopenvswitch : Check openvswitch containers ------------------------------ 2.92s\nopenvswitch : Flush Handlers -------------------------------------------- 2.16s\nopenvswitch : Ensuring config directories exist ------------------------- 1.85s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.47s\nmodule-load : Load modules ---------------------------------------------- 1.39s\nopenvswitch : include_tasks --------------------------------------------- 1.17s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 1.13s\nmodule-load : Drop module persistence ----------------------------------- 1.10s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.99s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 0.77s\nGroup hosts based on Kolla action --------------------------------------- 0.69s\nGroup hosts based on enabled services ----------------------------------- 0.57s\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state SUCCESS\u001b[0m\n\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nTuesday 13 January 2026  00:43:48 +0000 (0:00:00.170)       0:00:00.170 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nTuesday 13 January 2026  00:43:48 +0000 (0:00:00.686)       0:00:00.857 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nTuesday 13 January 2026  00:43:49 +0000 (0:00:00.508)       0:00:01.366 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nTuesday 13 January 2026  00:43:49 +0000 (0:00:00.591)       0:00:01.957 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nTuesday 13 January 2026  00:43:52 +0000 (0:00:02.678)       0:00:04.635 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nTuesday 13 January 2026  00:43:54 +0000 (0:00:01.815)       0:00:06.451 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:01.962)       0:00:08.414 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nTuesday 13 January 2026  00:43:57 +0000 (0:00:00.876)       0:00:09.290 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nTuesday 13 January 2026  00:43:58 +0000 (0:00:00.800)       0:00:10.090 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nTuesday 13 January 2026  00:43:58 +0000 (0:00:00.586)       0:00:10.677 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nTuesday 13 January 2026  00:43:59 +0000 (0:00:01.141)       0:00:11.818 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nTuesday 13 January 2026  00:44:00 +0000 (0:00:01.131)       0:00:12.950 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nTuesday 13 January 2026  00:44:06 +0000 (0:00:05.994)       0:00:18.944 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nTuesday 13 January 2026  00:44:08 +0000 (0:00:01.150)       0:00:20.095 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nTuesday 13 January 2026  00:44:10 +0000 (0:00:02.158)       0:00:22.253 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nTuesday 13 January 2026  00:44:11 +0000 (0:00:00.853)       0:00:23.106 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nTuesday 13 January 2026  00:44:12 +0000 (0:00:01.221)       0:00:24.327 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nTuesday 13 January 2026  00:44:13 +0000 (0:00:01.153)       0:00:25.481 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nTuesday 13 January 2026  00:44:15 +0000 (0:00:01.933)       0:00:27.414 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nTuesday 13 January 2026  00:44:16 +0000 (0:00:00.953)       0:00:28.367 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nTuesday 13 January 2026  00:44:17 +0000 (0:00:01.095)       0:00:29.463 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nTuesday 13 January 2026  00:44:18 +0000 (0:00:01.038)       0:00:30.502 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nTuesday 13 January 2026  00:44:19 +0000 (0:00:00.910)       0:00:31.413 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nTuesday 13 January 2026  00:44:19 +0000 (0:00:00.306)       0:00:31.719 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nTuesday 13 January 2026  00:44:20 +0000 (0:00:01.114)       0:00:32.833 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nTuesday 13 January 2026  00:44:22 +0000 (0:00:01.917)       0:00:34.751 ******* \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nTuesday 13 January 2026  00:44:23 +0000 (0:00:00.717)       0:00:35.469 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nTuesday 13 January 2026  00:44:26 +0000 (0:00:02.947)       0:00:38.417 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nTuesday 13 January 2026  00:44:26 +0000 (0:00:00.538)       0:00:38.955 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nTuesday 13 January 2026  00:44:28 +0000 (0:00:01.405)       0:00:40.361 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nTuesday 13 January 2026  00:44:30 +0000 (0:00:01.946)       0:00:42.308 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nTuesday 13 January 2026  00:44:30 +0000 (0:00:00.758)       0:00:43.066 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nTuesday 13 January 2026  00:44:31 +0000 (0:00:00.472)       0:00:43.539 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nTuesday 13 January 2026  00:44:33 +0000 (0:00:01.852)       0:00:45.391 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nTuesday 13 January 2026  00:44:36 +0000 (0:00:02.852)       0:00:48.244 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nTuesday 13 January 2026  00:44:36 +0000 (0:00:00.672)       0:00:48.916 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nTuesday 13 January 2026  00:45:20 +0000 (0:00:43.608)       0:01:32.525 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nTuesday 13 January 2026  00:45:20 +0000 (0:00:00.357)       0:01:32.883 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nTuesday 13 January 2026  00:45:21 +0000 (0:00:01.073)       0:01:33.957 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nTuesday 13 January 2026  00:45:23 +0000 (0:00:01.326)       0:01:35.284 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nTuesday 13 January 2026  00:45:48 +0000 (0:00:25.639)       0:02:00.923 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nTuesday 13 January 2026  00:45:49 +0000 (0:00:00.595)       0:02:01.519 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nTuesday 13 January 2026  00:45:50 +0000 (0:00:00.643)       0:02:02.162 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nTuesday 13 January 2026  00:45:50 +0000 (0:00:00.651)       0:02:02.813 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nTuesday 13 January 2026  00:45:51 +0000 (0:00:00.883)       0:02:03.697 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nTuesday 13 January 2026  00:45:51 +0000 (0:00:00.363)       0:02:04.060 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nTuesday 13 January 2026  00:45:52 +0000 (0:00:00.648)       0:02:04.708 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nTuesday 13 January 2026  00:45:53 +0000 (0:00:00.624)       0:02:05.333 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nTuesday 13 January 2026  00:45:54 +0000 (0:00:01.377)       0:02:06.711 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nTuesday 13 January 2026  00:45:55 +0000 (0:00:00.864)       0:02:07.576 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nTuesday 13 January 2026  00:45:55 +0000 (0:00:00.281)       0:02:07.857 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nTuesday 13 January 2026  00:45:56 +0000 (0:00:00.296)       0:02:08.153 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nTuesday 13 January 2026  00:45:56 +0000 (0:00:00.918)       0:02:09.072 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nTuesday 13 January 2026  00:45:57 +0000 (0:00:00.707)       0:02:09.779 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nTuesday 13 January 2026  00:46:00 +0000 (0:00:02.885)       0:02:12.665 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nTuesday 13 January 2026  00:46:01 +0000 (0:00:00.547)       0:02:13.213 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nTuesday 13 January 2026  00:46:01 +0000 (0:00:00.662)       0:02:13.875 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nTuesday 13 January 2026  00:46:02 +0000 (0:00:00.355)       0:02:14.231 ******* \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nTuesday 13 January 2026  00:46:02 +0000 (0:00:00.716)       0:02:14.947 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nTuesday 13 January 2026  00:46:03 +0000 (0:00:00.328)       0:02:15.276 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nTuesday 13 January 2026  00:46:03 +0000 (0:00:00.311)       0:02:15.587 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nTuesday 13 January 2026  00:46:03 +0000 (0:00:00.318)       0:02:15.906 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nTuesday 13 January 2026  00:46:04 +0000 (0:00:00.972)       0:02:16.878 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nTuesday 13 January 2026  00:46:05 +0000 (0:00:01.154)       0:02:18.032 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nTuesday 13 January 2026  00:46:07 +0000 (0:00:01.441)       0:02:19.474 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nTuesday 13 January 2026  00:46:19 +0000 (0:00:11.785)       0:02:31.260 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nTuesday 13 January 2026  00:46:19 +0000 (0:00:00.772)       0:02:32.033 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 13 January 2026  00:46:20 +0000 (0:00:00.517)       0:02:32.550 ******* \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 13 January 2026  00:46:20 +0000 (0:00:00.526)       0:02:33.077 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nTuesday 13 January 2026  00:46:21 +0000 (0:00:00.947)       0:02:34.024 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nTuesday 13 January 2026  00:46:22 +0000 (0:00:00.868)       0:02:34.892 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nTuesday 13 January 2026  00:46:24 +0000 (0:00:01.855)       0:02:36.748 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.915)       0:02:37.664 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.357)       0:02:38.022 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nTuesday 13 January 2026  00:46:26 +0000 (0:00:00.574)       0:02:38.596 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nTuesday 13 January 2026  00:46:26 +0000 (0:00:00.105)       0:02:38.702 ******* \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nTuesday 13 January 2026  00:46:26 +0000 (0:00:00.178)       0:02:38.880 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nTuesday 13 January 2026  00:46:27 +0000 (0:00:00.759)       0:02:39.639 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nTuesday 13 January 2026  00:46:28 +0000 (0:00:01.292)       0:02:40.932 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nTuesday 13 January 2026  00:46:29 +0000 (0:00:00.738)       0:02:41.670 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nTuesday 13 January 2026  00:46:29 +0000 (0:00:00.364)       0:02:42.035 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nTuesday 13 January 2026  00:46:36 +0000 (0:00:06.864)       0:02:48.899 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nTuesday 13 January 2026  00:46:48 +0000 (0:00:11.729)       0:03:00.628 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nTuesday 13 January 2026  00:46:49 +0000 (0:00:00.458)       0:03:01.087 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nTuesday 13 January 2026  00:46:49 +0000 (0:00:00.298)       0:03:01.386 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nTuesday 13 January 2026  00:46:49 +0000 (0:00:00.250)       0:03:01.636 ******* \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nTuesday 13 January 2026  00:46:50 +0000 (0:00:00.607)       0:03:02.243 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nTuesday 13 January 2026  00:46:51 +0000 (0:00:00.997)       0:03:03.241 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nTuesday 13 January 2026  00:46:51 +0000 (0:00:00.782)       0:03:04.023 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nTuesday 13 January 2026  00:46:52 +0000 (0:00:00.101)       0:03:04.125 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:00.985)       0:03:05.110 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:00.118)       0:03:05.228 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:00.116)       0:03:05.345 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:00.113)       0:03:05.459 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:00.129)       0:03:05.589 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nTuesday 13 January 2026  00:46:58 +0000 (0:00:05.335)       0:03:10.924 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nTuesday 13 January 2026  00:47:41 +0000 (0:00:42.203)       0:03:53.128 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nTuesday 13 January 2026  00:47:42 +0000 (0:00:01.205)       0:03:54.333 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nTuesday 13 January 2026  00:47:43 +0000 (0:00:01.557)       0:03:55.891 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nTuesday 13 January 2026  00:47:45 +0000 (0:00:01.209)       0:03:57.100 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nTuesday 13 January 2026  00:47:45 +0000 (0:00:00.101)       0:03:57.202 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nTuesday 13 January 2026  00:47:46 +0000 (0:00:01.545)       0:03:58.747 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.339)       0:03:59.086 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.952)       0:04:00.039 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nTuesday 13 January 2026  00:47:48 +0000 (0:00:00.131)       0:04:00.171 ******* \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nTuesday 13 January 2026  00:47:48 +0000 (0:00:00.195)       0:04:00.366 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nTuesday 13 January 2026  00:47:54 +0000 (0:00:06.380)       0:04:06.747 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nTuesday 13 January 2026  00:47:55 +0000 (0:00:01.215)       0:04:07.962 ******* \n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nTuesday 13 January 2026  00:48:07 +0000 (0:00:11.720)       0:04:19.682 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nTuesday 13 January 2026  00:48:08 +0000 (0:00:00.569)       0:04:20.252 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:48:08 +0000 (0:00:00.410)       0:04:20.662 ******* \n=============================================================================== \nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.61s\nk3s_server_post : Wait for Cilium resources ---------------------------- 42.20s\nk3s_server : Enable and check K3s service ------------------------------ 25.64s\nk3s_agent : Manage k3s service ----------------------------------------- 11.79s\nkubectl : Install required packages ------------------------------------ 11.73s\nManage labels ---------------------------------------------------------- 11.72s\nkubectl : Add repository Debian ----------------------------------------- 6.86s\nk9s : Install k9s packages ---------------------------------------------- 6.38s\nk3s_download : Download k3s binary x64 ---------------------------------- 5.99s\nk3s_server_post : Install Cilium ---------------------------------------- 5.34s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.95s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.89s\nk3s_server : Detect Kubernetes version for label compatibility ---------- 2.85s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.68s\nk3s_download : Download k3s binary armhf -------------------------------- 2.16s\nk3s_prereq : Enable IPv6 router advertisements -------------------------- 1.96s\nk3s_server : Copy vip manifest to first master -------------------------- 1.95s\nk3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 1.93s\nk3s_server : Create custom resolv.conf for k3s -------------------------- 1.92s\nMake kubeconfig available for use inside the manager service ------------ 1.86s\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy kubeconfig to the configuration repository] *************************\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 13 January 2026  00:48:13 +0000 (0:00:00.176)       0:00:00.176 ******* \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 13 January 2026  00:48:14 +0000 (0:00:00.718)       0:00:00.895 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig file] ****************************\nTuesday 13 January 2026  00:48:15 +0000 (0:00:00.994)       0:00:01.890 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:48:15 +0000 (0:00:00.413)       0:00:02.303 ******* \n=============================================================================== \nWrite kubeconfig file --------------------------------------------------- 0.99s\nGet kubeconfig file ----------------------------------------------------- 0.72s\nChange server address in the kubeconfig file ---------------------------- 0.41s\n\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nTuesday 13 January 2026  00:48:13 +0000 (0:00:00.151)       0:00:00.151 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nTuesday 13 January 2026  00:48:13 +0000 (0:00:00.474)       0:00:00.625 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 13 January 2026  00:48:14 +0000 (0:00:00.776)       0:00:01.402 ******* \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 13 January 2026  00:48:15 +0000 (0:00:00.690)       0:00:02.093 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nTuesday 13 January 2026  00:48:16 +0000 (0:00:01.165)       0:00:03.258 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nTuesday 13 January 2026  00:48:17 +0000 (0:00:00.477)       0:00:03.735 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nTuesday 13 January 2026  00:48:18 +0000 (0:00:01.412)       0:00:05.148 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:00.756)       0:00:05.905 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:00.414)       0:00:06.319 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:00.221)       0:00:06.541 ******* \n=============================================================================== \nMake kubeconfig available for use inside the manager service ------------ 1.41s\nWrite kubeconfig file --------------------------------------------------- 1.17s\nCreate .kube directory -------------------------------------------------- 0.78s\nChange server address in the kubeconfig inside the manager service ------ 0.76s\nGet kubeconfig file ----------------------------------------------------- 0.69s\nChange server address in the kubeconfig --------------------------------- 0.48s\nGet home directory of operator user ------------------------------------- 0.47s\nSet KUBECONFIG environment variable ------------------------------------- 0.41s\nEnable kubectl command line completion ---------------------------------- 0.22s\n\n\nPLAY [Set kolla_action_rabbitmq] ***********************************************\n\nTASK [Inform the user about the following task] ********************************\nTuesday 13 January 2026  00:46:30 +0000 (0:00:00.188)       0:00:00.188 ******* \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check RabbitMQ service] **************************************************\nTuesday 13 January 2026  00:46:30 +0000 (0:00:00.072)       0:00:00.261 ******* \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******\nTuesday 13 January 2026  00:46:33 +0000 (0:00:02.875)       0:00:03.136 ******* \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************\nTuesday 13 January 2026  00:46:33 +0000 (0:00:00.050)       0:00:03.186 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:46:33 +0000 (0:00:00.164)       0:00:03.351 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:46:33 +0000 (0:00:00.312)       0:00:03.664 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 13 January 2026  00:46:34 +0000 (0:00:00.518)       0:00:04.183 ******* \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nTuesday 13 January 2026  00:46:34 +0000 (0:00:00.562)       0:00:04.746 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nTuesday 13 January 2026  00:46:35 +0000 (0:00:00.989)       0:00:05.736 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nTuesday 13 January 2026  00:46:35 +0000 (0:00:00.387)       0:00:06.123 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nTuesday 13 January 2026  00:46:36 +0000 (0:00:00.555)       0:00:06.679 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nTuesday 13 January 2026  00:46:37 +0000 (0:00:00.492)       0:00:07.172 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 13 January 2026  00:46:37 +0000 (0:00:00.771)       0:00:07.943 ******* \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nTuesday 13 January 2026  00:46:38 +0000 (0:00:00.760)       0:00:08.704 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nTuesday 13 January 2026  00:46:39 +0000 (0:00:01.146)       0:00:09.850 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nTuesday 13 January 2026  00:46:40 +0000 (0:00:00.336)       0:00:10.186 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nTuesday 13 January 2026  00:46:40 +0000 (0:00:00.746)       0:00:10.933 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nTuesday 13 January 2026  00:46:41 +0000 (0:00:00.958)       0:00:11.891 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nTuesday 13 January 2026  00:46:43 +0000 (0:00:01.638)       0:00:13.529 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nTuesday 13 January 2026  00:46:45 +0000 (0:00:01.896)       0:00:15.426 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nTuesday 13 January 2026  00:46:47 +0000 (0:00:02.135)       0:00:17.562 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nTuesday 13 January 2026  00:46:49 +0000 (0:00:02.099)       0:00:19.662 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nTuesday 13 January 2026  00:46:51 +0000 (0:00:02.098)       0:00:21.760 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nTuesday 13 January 2026  00:46:53 +0000 (0:00:01.632)       0:00:23.393 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 13 January 2026  00:46:54 +0000 (0:00:01.696)       0:00:25.090 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Check rabbitmq containers] ************************************\nTuesday 13 January 2026  00:46:55 +0000 (0:00:00.504)       0:00:25.594 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Creating rabbitmq volume] *************************************\nTuesday 13 January 2026  00:46:57 +0000 (0:00:02.301)       0:00:27.896 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Running RabbitMQ bootstrap container] *************************\nTuesday 13 January 2026  00:46:59 +0000 (0:00:01.363)       0:00:29.259 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nTuesday 13 January 2026  00:47:07 +0000 (0:00:08.407)       0:00:37.666 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 13 January 2026  00:47:07 +0000 (0:00:00.250)       0:00:37.917 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 13 January 2026  00:47:08 +0000 (0:00:00.571)       0:00:38.489 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 13 January 2026  00:47:08 +0000 (0:00:00.202)       0:00:38.691 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 13 January 2026  00:47:10 +0000 (0:00:01.622)       0:00:40.313 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 13 January 2026  00:48:02 +0000 (0:00:51.987)       0:01:32.301 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 13 January 2026  00:48:02 +0000 (0:00:00.818)       0:01:33.119 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 13 January 2026  00:48:03 +0000 (0:00:00.277)       0:01:33.397 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 13 January 2026  00:48:10 +0000 (0:00:07.097)       0:01:40.494 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 13 January 2026  00:48:20 +0000 (0:00:10.583)       0:01:51.078 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 13 January 2026  00:48:21 +0000 (0:00:00.724)       0:01:51.803 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 13 January 2026  00:48:21 +0000 (0:00:00.186)       0:01:51.989 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 13 January 2026  00:48:23 +0000 (0:00:01.741)       0:01:53.731 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nTuesday 13 January 2026  00:48:37 +0000 (0:00:14.303)       0:02:08.034 ******* \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:00.518)       0:02:08.552 ******* \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35menable_outward_rabbitmq_True\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35moutward_rabbitmq_restart\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role rabbitmq (outward)] *******************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Restart rabbitmq (outward) services] *************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply rabbitmq (outward) post-configuration] *****************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:48:40 +0000 (0:00:02.145)       0:02:10.698 ******* \n=============================================================================== \nrabbitmq : Waiting for rabbitmq to start ------------------------------- 76.87s\nrabbitmq : Restart rabbitmq container ---------------------------------- 10.46s\nrabbitmq : Running RabbitMQ bootstrap container ------------------------- 8.41s\nCheck RabbitMQ service -------------------------------------------------- 2.88s\nrabbitmq : Check rabbitmq containers ------------------------------------ 2.30s\nrabbitmq : Enable all stable feature flags ------------------------------ 2.15s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.14s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 2.12s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 2.10s\nrabbitmq : Copying over advanced.config --------------------------------- 2.10s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.90s\nrabbitmq : Copying over enabled_plugins --------------------------------- 1.70s\nrabbitmq : Copying over config.json files for services ------------------ 1.64s\nrabbitmq : Copying over definitions.json -------------------------------- 1.63s\nrabbitmq : Creating rabbitmq volume ------------------------------------- 1.36s\nrabbitmq : Get container facts ------------------------------------------ 1.15s\nrabbitmq : Get container facts ------------------------------------------ 0.99s\nrabbitmq : Ensuring config directories exist ---------------------------- 0.96s\nrabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 0.77s\nrabbitmq : include_tasks ------------------------------------------------ 0.76s\n\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.147)       0:00:00.147 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:47:16 +0000 (0:00:00.820)       0:00:00.968 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.829)       0:00:01.797 ******* \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nTuesday 13 January 2026  00:47:18 +0000 (0:00:00.925)       0:00:02.723 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nTuesday 13 January 2026  00:47:19 +0000 (0:00:00.944)       0:00:03.667 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nTuesday 13 January 2026  00:47:20 +0000 (0:00:01.456)       0:00:05.124 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nTuesday 13 January 2026  00:47:21 +0000 (0:00:01.187)       0:00:06.311 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Check ovn-controller containers] ************************\nTuesday 13 January 2026  00:47:23 +0000 (0:00:02.037)       0:00:08.349 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nTuesday 13 January 2026  00:47:25 +0000 (0:00:01.860)       0:00:10.210 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nTuesday 13 January 2026  00:47:28 +0000 (0:00:02.695)       0:00:12.906 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:18.840)       0:00:31.746 ******* \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.050)       0:00:31.797 ******* \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.047)       0:00:31.844 ******* \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.053)       0:00:31.897 ******* \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.051)       0:00:31.949 ******* \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.049)       0:00:31.999 ******* \n\nRUNNING HANDLER [ovn-controller : Reload systemd config] ***********************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.049)       0:00:32.048 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nTuesday 13 January 2026  00:47:49 +0000 (0:00:02.017)       0:00:34.066 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 13 January 2026  00:48:18 +0000 (0:00:28.701)       0:01:02.768 ******* \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:00.706)       0:01:03.474 ******* \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:00.410)       0:01:03.885 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nTuesday 13 January 2026  00:48:20 +0000 (0:00:00.941)       0:01:04.826 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nTuesday 13 January 2026  00:48:20 +0000 (0:00:00.290)       0:01:05.116 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nTuesday 13 January 2026  00:48:20 +0000 (0:00:00.265)       0:01:05.382 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nTuesday 13 January 2026  00:48:21 +0000 (0:00:00.299)       0:01:05.681 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nTuesday 13 January 2026  00:48:21 +0000 (0:00:00.381)       0:01:06.063 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nTuesday 13 January 2026  00:48:21 +0000 (0:00:00.253)       0:01:06.317 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:00.265)       0:01:06.582 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:00.249)       0:01:06.832 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:00.351)       0:01:07.184 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:00.254)       0:01:07.439 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nTuesday 13 January 2026  00:48:23 +0000 (0:00:00.231)       0:01:07.670 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nTuesday 13 January 2026  00:48:23 +0000 (0:00:00.257)       0:01:07.928 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nTuesday 13 January 2026  00:48:23 +0000 (0:00:00.369)       0:01:08.297 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nTuesday 13 January 2026  00:48:24 +0000 (0:00:00.326)       0:01:08.624 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nTuesday 13 January 2026  00:48:24 +0000 (0:00:00.352)       0:01:08.976 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nTuesday 13 January 2026  00:48:24 +0000 (0:00:00.301)       0:01:09.278 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 13 January 2026  00:48:25 +0000 (0:00:00.343)       0:01:09.622 ******* \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************\nTuesday 13 January 2026  00:48:26 +0000 (0:00:00.913)       0:01:10.536 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************\nTuesday 13 January 2026  00:48:26 +0000 (0:00:00.466)       0:01:11.002 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check NB cluster status] ****************************************\nTuesday 13 January 2026  00:48:26 +0000 (0:00:00.452)       0:01:11.455 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check SB cluster status] ****************************************\nTuesday 13 January 2026  00:48:27 +0000 (0:00:00.586)       0:01:12.041 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***\nTuesday 13 January 2026  00:48:27 +0000 (0:00:00.337)       0:01:12.379 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***\nTuesday 13 January 2026  00:48:28 +0000 (0:00:00.330)       0:01:12.710 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************\nTuesday 13 January 2026  00:48:28 +0000 (0:00:00.351)       0:01:13.062 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************\nTuesday 13 January 2026  00:48:29 +0000 (0:00:00.568)       0:01:13.631 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  00:48:29 +0000 (0:00:00.299)       0:01:13.930 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nTuesday 13 January 2026  00:48:30 +0000 (0:00:01.425)       0:01:15.355 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nTuesday 13 January 2026  00:48:35 +0000 (0:00:04.484)       0:01:19.839 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:03.094)       0:01:22.934 ******* \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:00.069)       0:01:23.003 ******* \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:00.063)       0:01:23.067 ******* \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:00.067)       0:01:23.134 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nTuesday 13 January 2026  00:48:41 +0000 (0:00:02.377)       0:01:25.512 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nTuesday 13 January 2026  00:48:43 +0000 (0:00:02.393)       0:01:27.905 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nTuesday 13 January 2026  00:48:51 +0000 (0:00:07.858)       0:01:35.764 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nTuesday 13 January 2026  00:48:51 +0000 (0:00:00.131)       0:01:35.895 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nTuesday 13 January 2026  00:48:52 +0000 (0:00:00.709)       0:01:36.604 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nTuesday 13 January 2026  00:48:52 +0000 (0:00:00.524)       0:01:37.129 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nTuesday 13 January 2026  00:48:53 +0000 (0:00:00.687)       0:01:37.816 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nTuesday 13 January 2026  00:48:54 +0000 (0:00:00.828)       0:01:38.644 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nTuesday 13 January 2026  00:48:54 +0000 (0:00:00.792)       0:01:39.437 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Unset bootstrap args fact] **************************************\nTuesday 13 January 2026  00:48:55 +0000 (0:00:00.857)       0:01:40.295 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  00:48:56 +0000 (0:00:00.309)       0:01:40.604 ******* \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nTuesday 13 January 2026  00:48:57 +0000 (0:00:01.710)       0:01:42.314 ******* \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nTuesday 13 January 2026  00:49:02 +0000 (0:00:04.360)       0:01:46.675 ******* \n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:02.928)       0:01:49.604 ******* \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:00.060)       0:01:49.665 ******* \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:00.065)       0:01:49.731 ******* \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:00.063)       0:01:49.794 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nTuesday 13 January 2026  00:49:11 +0000 (0:00:06.304)       0:01:56.098 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nTuesday 13 January 2026  00:49:17 +0000 (0:00:06.211)       0:02:02.310 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nTuesday 13 January 2026  00:49:24 +0000 (0:00:06.433)       0:02:08.743 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nTuesday 13 January 2026  00:49:24 +0000 (0:00:00.144)       0:02:08.887 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nTuesday 13 January 2026  00:49:25 +0000 (0:00:00.893)       0:02:09.781 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nTuesday 13 January 2026  00:49:25 +0000 (0:00:00.663)       0:02:10.444 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nTuesday 13 January 2026  00:49:26 +0000 (0:00:00.788)       0:02:11.233 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nTuesday 13 January 2026  00:49:27 +0000 (0:00:00.612)       0:02:11.845 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nTuesday 13 January 2026  00:49:28 +0000 (0:00:00.787)       0:02:12.633 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:49:29 +0000 (0:00:00.910)       0:02:13.543 ******* \n=============================================================================== \novn-controller : Restart ovn-controller container ---------------------- 28.70s\novn-controller : Configure OVN in OVSDB -------------------------------- 18.84s\novn-db : Restart ovn-northd container ---------------------------------- 14.29s\novn-db : Restart ovn-nb-db container ------------------------------------ 8.68s\novn-db : Restart ovn-sb-db container ------------------------------------ 8.60s\novn-db : Copying over config.json files for services -------------------- 4.48s\novn-db : Copying over config.json files for services -------------------- 4.36s\novn-db : Check ovn containers ------------------------------------------- 3.09s\novn-db : Check ovn containers ------------------------------------------- 2.93s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.70s\novn-controller : Copying over systemd override -------------------------- 2.04s\novn-controller : Reload systemd config ---------------------------------- 2.02s\novn-controller : Check ovn-controller containers ------------------------ 1.86s\novn-db : Ensuring config directories exist ------------------------------ 1.71s\novn-controller : Copying over config.json files for services ------------ 1.46s\novn-db : Ensuring config directories exist ------------------------------ 1.43s\novn-controller : Ensuring systemd override directory exists ------------- 1.19s\novn-controller : Ensuring config directories exist ---------------------- 0.94s\novn-db : Checking for any existing OVN DB container volumes ------------- 0.94s\novn-controller : include_tasks ------------------------------------------ 0.93s\n\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.246)       0:00:00.246 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:46:08 +0000 (0:00:00.276)       0:00:00.522 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.376)       0:00:00.898 ******* \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nTuesday 13 January 2026  00:46:09 +0000 (0:00:00.528)       0:00:01.427 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nTuesday 13 January 2026  00:46:10 +0000 (0:00:00.639)       0:00:02.066 ******* \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nTuesday 13 January 2026  00:46:11 +0000 (0:00:00.886)       0:00:02.952 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nTuesday 13 January 2026  00:46:11 +0000 (0:00:00.621)       0:00:03.574 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 13 January 2026  00:46:15 +0000 (0:00:03.190)       0:00:06.765 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 13 January 2026  00:46:15 +0000 (0:00:00.707)       0:00:07.472 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 13 January 2026  00:46:17 +0000 (0:00:01.591)       0:00:09.063 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nTuesday 13 January 2026  00:46:18 +0000 (0:00:00.923)       0:00:09.987 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nTuesday 13 January 2026  00:46:20 +0000 (0:00:02.258)       0:00:12.246 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nTuesday 13 January 2026  00:46:21 +0000 (0:00:01.046)       0:00:13.292 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nTuesday 13 January 2026  00:46:23 +0000 (0:00:02.216)       0:00:15.509 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:01.379)       0:00:16.889 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nTuesday 13 January 2026  00:46:27 +0000 (0:00:01.835)       0:00:18.724 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nTuesday 13 January 2026  00:46:29 +0000 (0:00:02.425)       0:00:21.150 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nTuesday 13 January 2026  00:46:32 +0000 (0:00:02.965)       0:00:24.115 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nTuesday 13 January 2026  00:46:35 +0000 (0:00:03.274)       0:00:27.390 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nTuesday 13 January 2026  00:46:38 +0000 (0:00:02.272)       0:00:29.663 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nTuesday 13 January 2026  00:46:42 +0000 (0:00:04.447)       0:00:34.111 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nTuesday 13 January 2026  00:46:43 +0000 (0:00:00.516)       0:00:34.628 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nTuesday 13 January 2026  00:46:45 +0000 (0:00:02.745)       0:00:37.373 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nTuesday 13 January 2026  00:46:48 +0000 (0:00:02.891)       0:00:40.264 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nTuesday 13 January 2026  00:46:50 +0000 (0:00:01.654)       0:00:41.919 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nTuesday 13 January 2026  00:46:52 +0000 (0:00:01.865)       0:00:43.785 ******* \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***\nTuesday 13 January 2026  00:46:53 +0000 (0:00:01.120)       0:00:44.905 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:46:57 +0000 (0:00:03.844)       0:00:48.749 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***\nTuesday 13 January 2026  00:46:57 +0000 (0:00:00.678)       0:00:49.428 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nTuesday 13 January 2026  00:46:58 +0000 (0:00:00.904)       0:00:50.332 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:47:00 +0000 (0:00:01.622)       0:00:51.954 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  00:47:00 +0000 (0:00:00.547)       0:00:52.502 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******\nTuesday 13 January 2026  00:47:01 +0000 (0:00:00.874)       0:00:53.376 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:47:03 +0000 (0:00:01.217)       0:00:54.594 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****\nTuesday 13 January 2026  00:47:04 +0000 (0:00:01.146)       0:00:55.740 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nTuesday 13 January 2026  00:47:04 +0000 (0:00:00.792)       0:00:56.533 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nTuesday 13 January 2026  00:47:06 +0000 (0:00:01.731)       0:00:58.265 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nTuesday 13 January 2026  00:47:08 +0000 (0:00:01.417)       0:00:59.682 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Check loadbalancer containers] ****************************\nTuesday 13 January 2026  00:47:08 +0000 (0:00:00.809)       0:01:00.491 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nTuesday 13 January 2026  00:47:11 +0000 (0:00:02.483)       0:01:02.975 ******* \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nTuesday 13 January 2026  00:47:11 +0000 (0:00:00.523)       0:01:03.499 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nTuesday 13 January 2026  00:47:15 +0000 (0:00:04.074)       0:01:07.573 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nTuesday 13 January 2026  00:47:16 +0000 (0:00:00.884)       0:01:08.458 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.954)       0:01:09.413 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nTuesday 13 January 2026  00:47:18 +0000 (0:00:01.121)       0:01:10.534 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nTuesday 13 January 2026  00:47:20 +0000 (0:00:01.960)       0:01:12.495 ******* \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nTuesday 13 January 2026  00:47:21 +0000 (0:00:00.762)       0:01:13.257 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nTuesday 13 January 2026  00:47:26 +0000 (0:00:04.488)       0:01:17.745 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nTuesday 13 January 2026  00:47:26 +0000 (0:00:00.730)       0:01:18.476 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nTuesday 13 January 2026  00:47:27 +0000 (0:00:01.082)       0:01:19.559 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nTuesday 13 January 2026  00:47:29 +0000 (0:00:01.439)       0:01:20.999 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nTuesday 13 January 2026  00:47:32 +0000 (0:00:02.584)       0:01:23.583 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nTuesday 13 January 2026  00:47:32 +0000 (0:00:00.322)       0:01:23.906 ******* \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nTuesday 13 January 2026  00:47:33 +0000 (0:00:00.849)       0:01:24.755 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nTuesday 13 January 2026  00:47:35 +0000 (0:00:02.571)       0:01:27.327 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nTuesday 13 January 2026  00:47:37 +0000 (0:00:01.563)       0:01:28.890 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nTuesday 13 January 2026  00:47:39 +0000 (0:00:02.151)       0:01:31.042 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nTuesday 13 January 2026  00:47:40 +0000 (0:00:00.876)       0:01:31.919 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nTuesday 13 January 2026  00:47:41 +0000 (0:00:01.256)       0:01:33.175 ******* \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nTuesday 13 January 2026  00:47:42 +0000 (0:00:00.762)       0:01:33.937 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nTuesday 13 January 2026  00:47:46 +0000 (0:00:04.160)       0:01:38.098 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:01.041)       0:01:39.139 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nTuesday 13 January 2026  00:47:48 +0000 (0:00:01.073)       0:01:40.212 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nTuesday 13 January 2026  00:47:49 +0000 (0:00:01.203)       0:01:41.416 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nTuesday 13 January 2026  00:47:52 +0000 (0:00:02.680)       0:01:44.097 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nTuesday 13 January 2026  00:47:53 +0000 (0:00:00.704)       0:01:44.801 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nTuesday 13 January 2026  00:47:53 +0000 (0:00:00.411)       0:01:45.213 ******* \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nTuesday 13 January 2026  00:47:54 +0000 (0:00:00.853)       0:01:46.067 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nTuesday 13 January 2026  00:48:00 +0000 (0:00:06.153)       0:01:52.220 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nTuesday 13 January 2026  00:48:01 +0000 (0:00:01.179)       0:01:53.400 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nTuesday 13 January 2026  00:48:03 +0000 (0:00:01.414)       0:01:54.815 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nTuesday 13 January 2026  00:48:05 +0000 (0:00:01.809)       0:01:56.624 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nTuesday 13 January 2026  00:48:06 +0000 (0:00:01.834)       0:01:58.459 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nTuesday 13 January 2026  00:48:07 +0000 (0:00:00.479)       0:01:58.938 ******* \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nTuesday 13 January 2026  00:48:07 +0000 (0:00:00.614)       0:01:59.553 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nTuesday 13 January 2026  00:48:12 +0000 (0:00:04.617)       0:02:04.170 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nTuesday 13 January 2026  00:48:16 +0000 (0:00:03.548)       0:02:07.719 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nTuesday 13 January 2026  00:48:19 +0000 (0:00:03.101)       0:02:10.820 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nTuesday 13 January 2026  00:48:20 +0000 (0:00:01.342)       0:02:12.163 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:01.898)       0:02:14.061 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nTuesday 13 January 2026  00:48:22 +0000 (0:00:00.431)       0:02:14.493 ******* \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nTuesday 13 January 2026  00:48:23 +0000 (0:00:00.793)       0:02:15.287 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nTuesday 13 January 2026  00:48:27 +0000 (0:00:03.351)       0:02:18.639 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nTuesday 13 January 2026  00:48:27 +0000 (0:00:00.627)       0:02:19.266 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nTuesday 13 January 2026  00:48:28 +0000 (0:00:00.644)       0:02:19.911 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nTuesday 13 January 2026  00:48:29 +0000 (0:00:01.361)       0:02:21.272 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nTuesday 13 January 2026  00:48:31 +0000 (0:00:02.179)       0:02:23.452 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nTuesday 13 January 2026  00:48:32 +0000 (0:00:00.744)       0:02:24.196 ******* \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nTuesday 13 January 2026  00:48:33 +0000 (0:00:00.896)       0:02:25.093 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nTuesday 13 January 2026  00:48:37 +0000 (0:00:04.091)       0:02:29.185 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nTuesday 13 January 2026  00:48:38 +0000 (0:00:01.228)       0:02:30.413 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nTuesday 13 January 2026  00:48:39 +0000 (0:00:01.063)       0:02:31.476 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nTuesday 13 January 2026  00:48:41 +0000 (0:00:01.194)       0:02:32.671 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nTuesday 13 January 2026  00:48:43 +0000 (0:00:01.929)       0:02:34.600 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nTuesday 13 January 2026  00:48:43 +0000 (0:00:00.298)       0:02:34.899 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nTuesday 13 January 2026  00:48:43 +0000 (0:00:00.577)       0:02:35.476 ******* \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nTuesday 13 January 2026  00:48:44 +0000 (0:00:00.921)       0:02:36.398 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nTuesday 13 January 2026  00:48:48 +0000 (0:00:03.384)       0:02:39.783 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nTuesday 13 January 2026  00:48:49 +0000 (0:00:00.908)       0:02:40.691 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nTuesday 13 January 2026  00:48:49 +0000 (0:00:00.795)       0:02:41.487 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nTuesday 13 January 2026  00:48:51 +0000 (0:00:01.286)       0:02:42.773 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nTuesday 13 January 2026  00:48:53 +0000 (0:00:02.094)       0:02:44.868 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nTuesday 13 January 2026  00:48:53 +0000 (0:00:00.628)       0:02:45.497 ******* \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nTuesday 13 January 2026  00:48:54 +0000 (0:00:00.947)       0:02:46.444 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nTuesday 13 January 2026  00:48:58 +0000 (0:00:03.601)       0:02:50.045 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nTuesday 13 January 2026  00:48:59 +0000 (0:00:01.195)       0:02:51.241 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nTuesday 13 January 2026  00:49:00 +0000 (0:00:00.915)       0:02:52.157 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nTuesday 13 January 2026  00:49:01 +0000 (0:00:01.372)       0:02:53.529 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nTuesday 13 January 2026  00:49:04 +0000 (0:00:02.467)       0:02:55.997 ******* \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:01.247)       0:02:57.244 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nTuesday 13 January 2026  00:49:09 +0000 (0:00:03.583)       0:03:00.827 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nTuesday 13 January 2026  00:49:09 +0000 (0:00:00.588)       0:03:01.416 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nTuesday 13 January 2026  00:49:10 +0000 (0:00:01.012)       0:03:02.428 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nTuesday 13 January 2026  00:49:12 +0000 (0:00:01.257)       0:03:03.685 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nTuesday 13 January 2026  00:49:14 +0000 (0:00:02.105)       0:03:05.791 ******* \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nTuesday 13 January 2026  00:49:15 +0000 (0:00:01.405)       0:03:07.197 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nTuesday 13 January 2026  00:49:18 +0000 (0:00:03.093)       0:03:10.290 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nTuesday 13 January 2026  00:49:20 +0000 (0:00:02.048)       0:03:12.338 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nTuesday 13 January 2026  00:49:23 +0000 (0:00:02.305)       0:03:14.644 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nTuesday 13 January 2026  00:49:26 +0000 (0:00:02.972)       0:03:17.616 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nTuesday 13 January 2026  00:49:27 +0000 (0:00:01.883)       0:03:19.500 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nTuesday 13 January 2026  00:49:29 +0000 (0:00:01.412)       0:03:20.912 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nTuesday 13 January 2026  00:49:29 +0000 (0:00:00.338)       0:03:21.251 ******* \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nTuesday 13 January 2026  00:49:31 +0000 (0:00:01.415)       0:03:22.667 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nTuesday 13 January 2026  00:49:32 +0000 (0:00:01.701)       0:03:24.368 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nTuesday 13 January 2026  00:49:33 +0000 (0:00:00.445)       0:03:24.814 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nTuesday 13 January 2026  00:49:34 +0000 (0:00:00.844)       0:03:25.659 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nTuesday 13 January 2026  00:49:34 +0000 (0:00:00.447)       0:03:26.106 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nTuesday 13 January 2026  00:49:35 +0000 (0:00:01.252)       0:03:27.358 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nTuesday 13 January 2026  00:49:36 +0000 (0:00:00.320)       0:03:27.679 ******* \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nTuesday 13 January 2026  00:49:37 +0000 (0:00:01.431)       0:03:29.110 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nTuesday 13 January 2026  00:49:41 +0000 (0:00:04.155)       0:03:33.265 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nTuesday 13 January 2026  00:49:43 +0000 (0:00:01.410)       0:03:34.676 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nTuesday 13 January 2026  00:49:45 +0000 (0:00:02.092)       0:03:36.769 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nTuesday 13 January 2026  00:49:46 +0000 (0:00:01.139)       0:03:37.908 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nTuesday 13 January 2026  00:49:48 +0000 (0:00:02.021)       0:03:39.930 ******* \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nTuesday 13 January 2026  00:49:49 +0000 (0:00:01.162)       0:03:41.092 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nTuesday 13 January 2026  00:49:52 +0000 (0:00:03.483)       0:03:44.576 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nTuesday 13 January 2026  00:49:53 +0000 (0:00:00.498)       0:03:45.075 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nTuesday 13 January 2026  00:49:54 +0000 (0:00:00.729)       0:03:45.805 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nTuesday 13 January 2026  00:49:55 +0000 (0:00:01.186)       0:03:46.991 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nTuesday 13 January 2026  00:49:57 +0000 (0:00:02.087)       0:03:49.078 ******* \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nTuesday 13 January 2026  00:49:59 +0000 (0:00:01.512)       0:03:50.590 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nTuesday 13 January 2026  00:50:03 +0000 (0:00:04.378)       0:03:54.969 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nTuesday 13 January 2026  00:50:04 +0000 (0:00:01.396)       0:03:56.366 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nTuesday 13 January 2026  00:50:05 +0000 (0:00:00.917)       0:03:57.283 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nTuesday 13 January 2026  00:50:07 +0000 (0:00:01.535)       0:03:58.819 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nTuesday 13 January 2026  00:50:09 +0000 (0:00:02.129)       0:04:00.949 ******* \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nTuesday 13 January 2026  00:50:11 +0000 (0:00:01.708)       0:04:02.657 ******* \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nTuesday 13 January 2026  00:50:11 +0000 (0:00:00.882)       0:04:03.539 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nTuesday 13 January 2026  00:50:16 +0000 (0:00:04.297)       0:04:07.837 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nTuesday 13 January 2026  00:50:17 +0000 (0:00:01.392)       0:04:09.230 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 13 January 2026  00:50:19 +0000 (0:00:01.504)       0:04:10.734 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 13 January 2026  00:50:21 +0000 (0:00:02.320)       0:04:13.054 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nTuesday 13 January 2026  00:50:24 +0000 (0:00:02.972)       0:04:16.027 ******* \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nTuesday 13 January 2026  00:50:25 +0000 (0:00:01.430)       0:04:17.458 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nTuesday 13 January 2026  00:50:27 +0000 (0:00:01.251)       0:04:18.709 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nTuesday 13 January 2026  00:50:28 +0000 (0:00:01.358)       0:04:20.068 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 13 January 2026  00:50:30 +0000 (0:00:01.922)       0:04:21.990 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 13 January 2026  00:50:32 +0000 (0:00:02.541)       0:04:24.531 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nTuesday 13 January 2026  00:50:36 +0000 (0:00:03.284)       0:04:27.816 ******* \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nTuesday 13 January 2026  00:50:37 +0000 (0:00:00.926)       0:04:28.743 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nTuesday 13 January 2026  00:50:38 +0000 (0:00:01.304)       0:04:30.047 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nTuesday 13 January 2026  00:50:39 +0000 (0:00:01.339)       0:04:31.387 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 13 January 2026  00:50:41 +0000 (0:00:01.591)       0:04:32.979 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 13 January 2026  00:50:44 +0000 (0:00:02.648)       0:04:35.627 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nTuesday 13 January 2026  00:50:47 +0000 (0:00:03.544)       0:04:39.172 ******* \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nTuesday 13 January 2026  00:50:49 +0000 (0:00:01.611)       0:04:40.783 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nTuesday 13 January 2026  00:50:52 +0000 (0:00:03.531)       0:04:44.315 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nTuesday 13 January 2026  00:50:53 +0000 (0:00:00.728)       0:04:45.043 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nTuesday 13 January 2026  00:50:55 +0000 (0:00:01.584)       0:04:46.627 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nTuesday 13 January 2026  00:50:56 +0000 (0:00:01.333)       0:04:47.960 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nTuesday 13 January 2026  00:50:58 +0000 (0:00:02.055)       0:04:50.016 ******* \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nTuesday 13 January 2026  00:50:59 +0000 (0:00:01.348)       0:04:51.365 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nTuesday 13 January 2026  00:51:05 +0000 (0:00:06.000)       0:04:57.366 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nTuesday 13 January 2026  00:51:06 +0000 (0:00:00.639)       0:04:58.005 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nTuesday 13 January 2026  00:51:07 +0000 (0:00:00.942)       0:04:58.948 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nTuesday 13 January 2026  00:51:08 +0000 (0:00:00.838)       0:04:59.787 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nTuesday 13 January 2026  00:51:09 +0000 (0:00:01.407)       0:05:01.194 ******* \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nTuesday 13 January 2026  00:51:10 +0000 (0:00:01.382)       0:05:02.576 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nTuesday 13 January 2026  00:51:15 +0000 (0:00:04.524)       0:05:07.101 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nTuesday 13 January 2026  00:51:16 +0000 (0:00:01.271)       0:05:08.372 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nTuesday 13 January 2026  00:51:17 +0000 (0:00:00.971)       0:05:09.344 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nTuesday 13 January 2026  00:51:18 +0000 (0:00:00.456)       0:05:09.801 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nTuesday 13 January 2026  00:51:19 +0000 (0:00:01.569)       0:05:11.370 ******* \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nTuesday 13 January 2026  00:51:21 +0000 (0:00:01.857)       0:05:13.228 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nTuesday 13 January 2026  00:51:24 +0000 (0:00:02.556)       0:05:15.785 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nTuesday 13 January 2026  00:51:24 +0000 (0:00:00.397)       0:05:16.183 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nTuesday 13 January 2026  00:51:25 +0000 (0:00:00.981)       0:05:17.164 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nTuesday 13 January 2026  00:51:26 +0000 (0:00:00.428)       0:05:17.593 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nTuesday 13 January 2026  00:51:27 +0000 (0:00:01.307)       0:05:18.901 ******* \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nTuesday 13 January 2026  00:51:29 +0000 (0:00:01.742)       0:05:20.644 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nTuesday 13 January 2026  00:51:35 +0000 (0:00:06.315)       0:05:26.960 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nTuesday 13 January 2026  00:51:35 +0000 (0:00:00.615)       0:05:27.575 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nTuesday 13 January 2026  00:51:37 +0000 (0:00:01.639)       0:05:29.215 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nTuesday 13 January 2026  00:51:38 +0000 (0:00:01.351)       0:05:30.566 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : swift] ****************************************************\nTuesday 13 January 2026  00:51:41 +0000 (0:00:02.165)       0:05:32.732 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nTuesday 13 January 2026  00:51:41 +0000 (0:00:00.323)       0:05:33.056 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nTuesday 13 January 2026  00:51:41 +0000 (0:00:00.311)       0:05:33.367 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nTuesday 13 January 2026  00:51:42 +0000 (0:00:00.637)       0:05:34.005 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nTuesday 13 January 2026  00:51:42 +0000 (0:00:00.328)       0:05:34.333 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nTuesday 13 January 2026  00:51:43 +0000 (0:00:00.312)       0:05:34.646 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nTuesday 13 January 2026  00:51:43 +0000 (0:00:00.844)       0:05:35.490 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nTuesday 13 January 2026  00:51:44 +0000 (0:00:00.699)       0:05:36.190 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nTuesday 13 January 2026  00:51:44 +0000 (0:00:00.359)       0:05:36.550 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nTuesday 13 January 2026  00:51:45 +0000 (0:00:00.895)       0:05:37.446 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nTuesday 13 January 2026  00:51:47 +0000 (0:00:01.336)       0:05:38.782 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nTuesday 13 January 2026  00:51:48 +0000 (0:00:00.960)       0:05:39.743 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nTuesday 13 January 2026  00:51:57 +0000 (0:00:09.590)       0:05:49.334 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nTuesday 13 January 2026  00:51:58 +0000 (0:00:00.816)       0:05:50.151 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nTuesday 13 January 2026  00:52:07 +0000 (0:00:08.751)       0:05:58.902 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nTuesday 13 January 2026  00:52:11 +0000 (0:00:04.183)       0:06:03.085 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nTuesday 13 January 2026  00:52:15 +0000 (0:00:04.425)       0:06:07.511 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nTuesday 13 January 2026  00:52:16 +0000 (0:00:00.367)       0:06:07.878 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nTuesday 13 January 2026  00:52:16 +0000 (0:00:00.356)       0:06:08.235 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nTuesday 13 January 2026  00:52:17 +0000 (0:00:00.736)       0:06:08.972 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nTuesday 13 January 2026  00:52:17 +0000 (0:00:00.334)       0:06:09.306 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nTuesday 13 January 2026  00:52:18 +0000 (0:00:00.379)       0:06:09.686 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nTuesday 13 January 2026  00:52:18 +0000 (0:00:00.339)       0:06:10.025 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:05.138)       0:06:15.163 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:52:24 +0000 (0:00:00.907)       0:06:16.071 ******* \n=============================================================================== \nloadbalancer : Start backup haproxy container --------------------------- 9.59s\nloadbalancer : Start backup proxysql container -------------------------- 8.75s\nhaproxy-config : Copying over skyline haproxy config -------------------- 6.32s\nhaproxy-config : Copying over designate haproxy config ------------------ 6.15s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 6.00s\nloadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.14s\nhaproxy-config : Copying over glance haproxy config --------------------- 4.62s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 4.52s\nhaproxy-config : Copying over barbican haproxy config ------------------- 4.49s\nloadbalancer : Copying over proxysql config ----------------------------- 4.45s\nloadbalancer : Start backup keepalived container ------------------------ 4.43s\nhaproxy-config : Copying over nova haproxy config ----------------------- 4.38s\nhaproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.30s\nloadbalancer : Wait for backup proxysql to start ------------------------ 4.18s\nhaproxy-config : Copying over cinder haproxy config --------------------- 4.16s\nhaproxy-config : Copying over neutron haproxy config -------------------- 4.16s\nhaproxy-config : Copying over horizon haproxy config -------------------- 4.09s\nhaproxy-config : Copying over aodh haproxy config ----------------------- 4.07s\nservice-cert-copy : loadbalancer | Copying over extra CA certificates --- 3.84s\nhaproxy-config : Copying over magnum haproxy config --------------------- 3.60s\n\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:52:29 +0000 (0:00:00.287)       0:00:00.287 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:52:30 +0000 (0:00:00.286)       0:00:00.574 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 13 January 2026  00:52:30 +0000 (0:00:00.456)       0:00:01.031 ******* \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nTuesday 13 January 2026  00:52:31 +0000 (0:00:00.475)       0:00:01.507 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nTuesday 13 January 2026  00:52:31 +0000 (0:00:00.707)       0:00:02.214 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 13 January 2026  00:52:34 +0000 (0:00:02.228)       0:00:04.442 ******* \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nTuesday 13 January 2026  00:52:34 +0000 (0:00:00.574)       0:00:05.017 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:52:37 +0000 (0:00:03.053)       0:00:08.071 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nTuesday 13 January 2026  00:52:38 +0000 (0:00:00.996)       0:00:09.067 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nTuesday 13 January 2026  00:52:39 +0000 (0:00:00.818)       0:00:09.886 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nTuesday 13 January 2026  00:52:41 +0000 (0:00:02.177)       0:00:12.064 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nTuesday 13 January 2026  00:52:44 +0000 (0:00:03.351)       0:00:15.415 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Check opensearch containers] ********************************\nTuesday 13 January 2026  00:52:46 +0000 (0:00:01.693)       0:00:17.109 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 13 January 2026  00:52:48 +0000 (0:00:02.152)       0:00:19.262 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.256)       0:00:19.519 ******* \n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.058)       0:00:19.577 ******* \n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.058)       0:00:19.636 ******* \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.060)       0:00:19.696 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.191)       0:00:19.888 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.457)       0:00:20.346 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nTuesday 13 January 2026  00:53:46 +0000 (0:00:56.663)       0:01:17.009 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 13 January 2026  00:54:50 +0000 (0:01:03.539)       0:02:20.549 ******* \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nTuesday 13 January 2026  00:54:50 +0000 (0:00:00.709)       0:02:21.258 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch cluster to become healthy] **************\nTuesday 13 January 2026  00:54:53 +0000 (0:00:02.488)       0:02:23.746 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nTuesday 13 January 2026  00:54:55 +0000 (0:00:02.171)       0:02:25.918 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nTuesday 13 January 2026  00:54:58 +0000 (0:00:02.612)       0:02:28.531 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nTuesday 13 January 2026  00:55:01 +0000 (0:00:03.225)       0:02:31.756 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:55:03 +0000 (0:00:02.498)       0:02:34.255 ******* \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 63.54s\nopensearch : Restart opensearch container ------------------------------ 56.66s\nopensearch : Copying over opensearch service config file ---------------- 3.35s\nopensearch : Create new log retention policy ---------------------------- 3.23s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 3.05s\nopensearch : Check if a log retention policy exists --------------------- 2.61s\nopensearch : Apply retention policy to existing indices ----------------- 2.50s\nopensearch : Wait for OpenSearch to become ready ------------------------ 2.49s\nopensearch : Ensuring config directories exist -------------------------- 2.23s\nopensearch : Copying over config.json files for services ---------------- 2.18s\nopensearch : Wait for OpenSearch cluster to become healthy -------------- 2.17s\nopensearch : Check opensearch containers -------------------------------- 2.15s\nopensearch : Copying over opensearch-dashboards config file ------------- 1.69s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.00s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 0.82s\nopensearch : include_tasks ---------------------------------------------- 0.71s\nopensearch : Setting sysctl values -------------------------------------- 0.71s\nopensearch : include_tasks ---------------------------------------------- 0.57s\nopensearch : include_tasks ---------------------------------------------- 0.48s\nopensearch : Perform a flush -------------------------------------------- 0.46s\n\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Prepare deployment of Ceph services] *************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nTuesday 13 January 2026  00:43:48 +0000 (0:00:00.666)       0:00:00.666 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nTuesday 13 January 2026  00:43:49 +0000 (0:00:01.021)       0:00:01.687 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nTuesday 13 January 2026  00:43:51 +0000 (0:00:01.566)       0:00:03.254 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nTuesday 13 January 2026  00:43:51 +0000 (0:00:00.735)       0:00:03.990 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nTuesday 13 January 2026  00:43:52 +0000 (0:00:00.897)       0:00:04.887 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nTuesday 13 January 2026  00:43:53 +0000 (0:00:00.660)       0:00:05.548 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nTuesday 13 January 2026  00:43:53 +0000 (0:00:00.516)       0:00:06.065 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nTuesday 13 January 2026  00:43:54 +0000 (0:00:00.809)       0:00:06.874 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nTuesday 13 January 2026  00:43:55 +0000 (0:00:00.680)       0:00:07.555 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:00.641)       0:00:08.197 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nTuesday 13 January 2026  00:43:56 +0000 (0:00:00.642)       0:00:08.839 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nTuesday 13 January 2026  00:43:57 +0000 (0:00:01.208)       0:00:10.048 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nTuesday 13 January 2026  00:44:01 +0000 (0:00:03.342)       0:00:13.391 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nTuesday 13 January 2026  00:44:02 +0000 (0:00:00.772)       0:00:14.163 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nTuesday 13 January 2026  00:44:03 +0000 (0:00:01.044)       0:00:15.208 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nTuesday 13 January 2026  00:44:03 +0000 (0:00:00.405)       0:00:15.614 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-01-13 00:43:58.698129', 'end': '2026-01-13 00:43:59.016841', 'delta': '0:00:00.318712', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-01-13 00:43:59.859242', 'end': '2026-01-13 00:44:00.158616', 'delta': '0:00:00.299374', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-01-13 00:44:00.795375', 'end': '2026-01-13 00:44:01.103968', 'delta': '0:00:00.308593', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nTuesday 13 January 2026  00:44:03 +0000 (0:00:00.285)       0:00:15.899 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nTuesday 13 January 2026  00:44:05 +0000 (0:00:01.664)       0:00:17.564 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nTuesday 13 January 2026  00:44:06 +0000 (0:00:00.757)       0:00:18.322 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nTuesday 13 January 2026  00:44:08 +0000 (0:00:02.089)       0:00:20.412 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 13 January 2026  00:44:10 +0000 (0:00:01.944)       0:00:22.356 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nTuesday 13 January 2026  00:44:11 +0000 (0:00:01.503)       0:00:23.859 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nTuesday 13 January 2026  00:44:11 +0000 (0:00:00.126)       0:00:23.986 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 13 January 2026  00:44:12 +0000 (0:00:00.377)       0:00:24.364 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nTuesday 13 January 2026  00:44:13 +0000 (0:00:01.133)       0:00:25.497 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nTuesday 13 January 2026  00:44:14 +0000 (0:00:01.175)       0:00:26.672 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nTuesday 13 January 2026  00:44:15 +0000 (0:00:00.680)       0:00:27.352 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nTuesday 13 January 2026  00:44:16 +0000 (0:00:01.590)       0:00:28.943 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nTuesday 13 January 2026  00:44:17 +0000 (0:00:00.743)       0:00:29.686 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nTuesday 13 January 2026  00:44:18 +0000 (0:00:00.825)       0:00:30.512 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nTuesday 13 January 2026  00:44:19 +0000 (0:00:00.678)       0:00:31.190 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-02-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-00-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nTuesday 13 January 2026  00:44:20 +0000 (0:00:01.230)       0:00:32.421 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-02-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-00-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nTuesday 13 January 2026  00:44:21 +0000 (0:00:01.434)       0:00:33.855 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nTuesday 13 January 2026  00:44:23 +0000 (0:00:01.356)       0:00:35.212 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 13 January 2026  00:44:23 +0000 (0:00:00.548)       0:00:35.761 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 13 January 2026  00:44:24 +0000 (0:00:01.247)       0:00:37.009 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 13 January 2026  00:44:25 +0000 (0:00:00.834)       0:00:37.843 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 13 January 2026  00:44:26 +0000 (0:00:00.916)       0:00:38.760 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nTuesday 13 January 2026  00:44:27 +0000 (0:00:00.852)       0:00:39.613 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nTuesday 13 January 2026  00:44:32 +0000 (0:00:05.016)       0:00:44.630 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nTuesday 13 January 2026  00:44:33 +0000 (0:00:01.108)       0:00:45.738 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 13 January 2026  00:44:34 +0000 (0:00:01.119)       0:00:46.858 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 13 January 2026  00:44:35 +0000 (0:00:00.611)       0:00:47.469 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 13 January 2026  00:44:36 +0000 (0:00:00.647)       0:00:48.117 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 13 January 2026  00:44:36 +0000 (0:00:00.596)       0:00:48.713 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 13 January 2026  00:44:37 +0000 (0:00:00.847)       0:00:49.561 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 13 January 2026  00:44:37 +0000 (0:00:00.328)       0:00:49.889 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 13 January 2026  00:44:38 +0000 (0:00:00.349)       0:00:50.239 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 13 January 2026  00:44:38 +0000 (0:00:00.477)       0:00:50.716 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 13 January 2026  00:44:39 +0000 (0:00:00.382)       0:00:51.099 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nTuesday 13 January 2026  00:44:39 +0000 (0:00:00.883)       0:00:51.982 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nTuesday 13 January 2026  00:44:40 +0000 (0:00:00.984)       0:00:52.967 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:44:43 +0000 (0:00:02.195)       0:00:55.162 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:44:44 +0000 (0:00:01.695)       0:00:56.858 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:44:46 +0000 (0:00:01.648)       0:00:58.506 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:44:48 +0000 (0:00:01.632)       0:01:00.139 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:44:49 +0000 (0:00:01.005)       0:01:01.144 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:44:50 +0000 (0:00:01.056)       0:01:02.200 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:44:50 +0000 (0:00:00.852)       0:01:03.053 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:44:52 +0000 (0:00:01.470)       0:01:04.523 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:44:52 +0000 (0:00:00.544)       0:01:05.067 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:44:53 +0000 (0:00:00.802)       0:01:05.870 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:44:54 +0000 (0:00:01.099)       0:01:06.970 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:44:56 +0000 (0:00:01.579)       0:01:08.549 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:44:57 +0000 (0:00:00.635)       0:01:09.185 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:44:58 +0000 (0:00:01.001)       0:01:10.186 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:44:58 +0000 (0:00:00.630)       0:01:10.817 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:44:59 +0000 (0:00:00.921)       0:01:11.739 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:45:00 +0000 (0:00:00.679)       0:01:12.418 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:45:01 +0000 (0:00:01.113)       0:01:13.532 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:45:02 +0000 (0:00:00.854)       0:01:14.386 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:45:03 +0000 (0:00:01.465)       0:01:15.851 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:45:04 +0000 (0:00:00.649)       0:01:16.500 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nTuesday 13 January 2026  00:45:05 +0000 (0:00:01.315)       0:01:17.816 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nTuesday 13 January 2026  00:45:07 +0000 (0:00:01.946)       0:01:19.762 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nTuesday 13 January 2026  00:45:10 +0000 (0:00:03.006)       0:01:22.768 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nTuesday 13 January 2026  00:45:11 +0000 (0:00:01.013)       0:01:23.781 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nTuesday 13 January 2026  00:45:12 +0000 (0:00:00.648)       0:01:24.430 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nTuesday 13 January 2026  00:45:13 +0000 (0:00:00.779)       0:01:25.210 ******* \n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nTuesday 13 January 2026  00:45:14 +0000 (0:00:01.510)       0:01:26.720 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nTuesday 13 January 2026  00:45:16 +0000 (0:00:02.169)       0:01:28.889 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nTuesday 13 January 2026  00:45:17 +0000 (0:00:01.037)       0:01:29.927 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nTuesday 13 January 2026  00:45:18 +0000 (0:00:00.968)       0:01:30.895 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nTuesday 13 January 2026  00:45:19 +0000 (0:00:00.561)       0:01:31.457 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nTuesday 13 January 2026  00:45:20 +0000 (0:00:01.058)       0:01:32.516 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nTuesday 13 January 2026  00:46:21 +0000 (0:01:00.887)       0:02:33.404 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nTuesday 13 January 2026  00:46:22 +0000 (0:00:00.950)       0:02:34.354 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nTuesday 13 January 2026  00:46:23 +0000 (0:00:00.968)       0:02:35.323 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nTuesday 13 January 2026  00:46:23 +0000 (0:00:00.147)       0:02:35.471 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nTuesday 13 January 2026  00:46:23 +0000 (0:00:00.606)       0:02:36.077 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nTuesday 13 January 2026  00:46:24 +0000 (0:00:00.786)       0:02:36.864 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nTuesday 13 January 2026  00:46:25 +0000 (0:00:00.590)       0:02:37.455 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nTuesday 13 January 2026  00:46:27 +0000 (0:00:02.165)       0:02:39.620 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nTuesday 13 January 2026  00:46:27 +0000 (0:00:00.472)       0:02:40.093 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nTuesday 13 January 2026  00:46:28 +0000 (0:00:00.979)       0:02:41.072 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nTuesday 13 January 2026  00:46:29 +0000 (0:00:00.754)       0:02:41.827 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nTuesday 13 January 2026  00:46:30 +0000 (0:00:00.597)       0:02:42.424 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nTuesday 13 January 2026  00:46:30 +0000 (0:00:00.645)       0:02:43.070 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nTuesday 13 January 2026  00:46:31 +0000 (0:00:00.645)       0:02:43.715 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nTuesday 13 January 2026  00:46:32 +0000 (0:00:00.679)       0:02:44.395 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nTuesday 13 January 2026  00:46:32 +0000 (0:00:00.652)       0:02:45.047 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nTuesday 13 January 2026  00:46:33 +0000 (0:00:00.710)       0:02:45.757 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nTuesday 13 January 2026  00:46:34 +0000 (0:00:00.558)       0:02:46.316 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nTuesday 13 January 2026  00:46:35 +0000 (0:00:01.010)       0:02:47.326 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nTuesday 13 January 2026  00:46:36 +0000 (0:00:01.087)       0:02:48.414 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nTuesday 13 January 2026  00:46:43 +0000 (0:00:07.433)       0:02:55.847 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nTuesday 13 January 2026  00:46:44 +0000 (0:00:00.840)       0:02:56.687 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nTuesday 13 January 2026  00:46:45 +0000 (0:00:00.844)       0:02:57.532 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nTuesday 13 January 2026  00:46:46 +0000 (0:00:01.323)       0:02:58.855 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nTuesday 13 January 2026  00:46:47 +0000 (0:00:00.569)       0:02:59.425 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nTuesday 13 January 2026  00:46:48 +0000 (0:00:00.745)       0:03:00.170 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nTuesday 13 January 2026  00:46:48 +0000 (0:00:00.603)       0:03:00.774 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nTuesday 13 January 2026  00:46:49 +0000 (0:00:00.736)       0:03:01.510 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nTuesday 13 January 2026  00:46:50 +0000 (0:00:00.626)       0:03:02.137 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nTuesday 13 January 2026  00:46:51 +0000 (0:00:01.146)       0:03:03.283 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nTuesday 13 January 2026  00:46:51 +0000 (0:00:00.512)       0:03:03.795 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nTuesday 13 January 2026  00:46:52 +0000 (0:00:00.793)       0:03:04.589 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nTuesday 13 January 2026  00:46:55 +0000 (0:00:02.915)       0:03:07.504 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nTuesday 13 January 2026  00:46:56 +0000 (0:00:01.085)       0:03:08.590 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nTuesday 13 January 2026  00:46:57 +0000 (0:00:00.843)       0:03:09.433 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nTuesday 13 January 2026  00:46:58 +0000 (0:00:01.148)       0:03:10.582 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nTuesday 13 January 2026  00:46:59 +0000 (0:00:00.653)       0:03:11.235 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nTuesday 13 January 2026  00:46:59 +0000 (0:00:00.811)       0:03:12.047 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nTuesday 13 January 2026  00:47:00 +0000 (0:00:00.499)       0:03:12.546 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 13 January 2026  00:47:01 +0000 (0:00:00.698)       0:03:13.245 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 13 January 2026  00:47:01 +0000 (0:00:00.632)       0:03:13.877 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 13 January 2026  00:47:02 +0000 (0:00:00.786)       0:03:14.663 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 13 January 2026  00:47:03 +0000 (0:00:00.527)       0:03:15.191 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 13 January 2026  00:47:03 +0000 (0:00:00.756)       0:03:15.947 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 13 January 2026  00:47:04 +0000 (0:00:00.359)       0:03:16.307 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 13 January 2026  00:47:04 +0000 (0:00:00.356)       0:03:16.663 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 13 January 2026  00:47:04 +0000 (0:00:00.337)       0:03:17.000 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 13 January 2026  00:47:05 +0000 (0:00:00.500)       0:03:17.501 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nTuesday 13 January 2026  00:47:06 +0000 (0:00:01.414)       0:03:18.916 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:47:08 +0000 (0:00:02.179)       0:03:21.095 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nTuesday 13 January 2026  00:47:09 +0000 (0:00:00.964)       0:03:22.060 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nTuesday 13 January 2026  00:47:10 +0000 (0:00:00.851)       0:03:22.911 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nTuesday 13 January 2026  00:47:11 +0000 (0:00:00.269)       0:03:23.180 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nTuesday 13 January 2026  00:47:12 +0000 (0:00:01.317)       0:03:24.498 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nTuesday 13 January 2026  00:47:13 +0000 (0:00:00.629)       0:03:25.127 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nTuesday 13 January 2026  00:47:13 +0000 (0:00:00.360)       0:03:25.488 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nTuesday 13 January 2026  00:47:14 +0000 (0:00:01.080)       0:03:26.568 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nTuesday 13 January 2026  00:47:14 +0000 (0:00:00.386)       0:03:26.955 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.310)       0:03:27.266 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.193)       0:03:27.459 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.252)       0:03:27.712 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.181)       0:03:27.893 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nTuesday 13 January 2026  00:47:15 +0000 (0:00:00.195)       0:03:28.089 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nTuesday 13 January 2026  00:47:16 +0000 (0:00:00.110)       0:03:28.199 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nTuesday 13 January 2026  00:47:16 +0000 (0:00:00.530)       0:03:28.729 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nTuesday 13 January 2026  00:47:16 +0000 (0:00:00.175)       0:03:28.905 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.378)       0:03:29.283 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.295)       0:03:29.579 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.174)       0:03:29.754 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nTuesday 13 January 2026  00:47:17 +0000 (0:00:00.201)       0:03:29.956 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nTuesday 13 January 2026  00:47:18 +0000 (0:00:00.854)       0:03:30.810 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nTuesday 13 January 2026  00:47:18 +0000 (0:00:00.260)       0:03:31.070 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nTuesday 13 January 2026  00:47:20 +0000 (0:00:01.208)       0:03:32.278 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nTuesday 13 January 2026  00:47:20 +0000 (0:00:00.698)       0:03:32.976 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nTuesday 13 January 2026  00:47:21 +0000 (0:00:00.413)       0:03:33.389 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nTuesday 13 January 2026  00:47:21 +0000 (0:00:00.690)       0:03:34.080 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nTuesday 13 January 2026  00:47:22 +0000 (0:00:00.631)       0:03:34.711 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nTuesday 13 January 2026  00:47:23 +0000 (0:00:01.327)       0:03:36.039 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nTuesday 13 January 2026  00:47:24 +0000 (0:00:00.712)       0:03:36.752 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nTuesday 13 January 2026  00:47:25 +0000 (0:00:00.355)       0:03:37.108 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nTuesday 13 January 2026  00:47:25 +0000 (0:00:00.911)       0:03:38.020 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nTuesday 13 January 2026  00:47:26 +0000 (0:00:00.857)       0:03:38.878 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nTuesday 13 January 2026  00:47:27 +0000 (0:00:00.545)       0:03:39.424 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nTuesday 13 January 2026  00:47:28 +0000 (0:00:01.170)       0:03:40.594 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nTuesday 13 January 2026  00:47:29 +0000 (0:00:00.643)       0:03:41.238 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mon] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:47:29 +0000 (0:00:00.644)       0:03:41.882 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:47:30 +0000 (0:00:00.954)       0:03:42.837 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:47:31 +0000 (0:00:00.512)       0:03:43.349 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:47:32 +0000 (0:00:01.104)       0:03:44.454 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:47:32 +0000 (0:00:00.324)       0:03:44.779 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:47:33 +0000 (0:00:00.325)       0:03:45.104 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:47:33 +0000 (0:00:00.330)       0:03:45.435 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:47:34 +0000 (0:00:01.105)       0:03:46.540 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:47:34 +0000 (0:00:00.313)       0:03:46.854 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:47:35 +0000 (0:00:00.333)       0:03:47.187 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:47:35 +0000 (0:00:00.649)       0:03:47.837 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:47:36 +0000 (0:00:01.059)       0:03:48.896 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:47:37 +0000 (0:00:00.318)       0:03:49.214 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:47:37 +0000 (0:00:00.305)       0:03:49.520 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:47:37 +0000 (0:00:00.324)       0:03:49.844 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:47:38 +0000 (0:00:00.365)       0:03:50.210 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:47:38 +0000 (0:00:00.630)       0:03:50.840 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:47:39 +0000 (0:00:00.452)       0:03:51.293 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:47:39 +0000 (0:00:00.338)       0:03:51.632 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:47:39 +0000 (0:00:00.334)       0:03:51.966 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:47:40 +0000 (0:00:00.718)       0:03:52.684 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nTuesday 13 January 2026  00:47:41 +0000 (0:00:00.538)       0:03:53.223 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nTuesday 13 January 2026  00:47:41 +0000 (0:00:00.340)       0:03:53.563 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nTuesday 13 January 2026  00:47:42 +0000 (0:00:00.923)       0:03:54.486 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nTuesday 13 January 2026  00:47:42 +0000 (0:00:00.164)       0:03:54.651 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nTuesday 13 January 2026  00:47:43 +0000 (0:00:01.169)       0:03:55.821 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nTuesday 13 January 2026  00:47:44 +0000 (0:00:00.482)       0:03:56.304 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nTuesday 13 January 2026  00:47:44 +0000 (0:00:00.444)       0:03:56.748 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nTuesday 13 January 2026  00:47:45 +0000 (0:00:01.189)       0:03:57.937 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nTuesday 13 January 2026  00:47:46 +0000 (0:00:00.597)       0:03:58.535 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nTuesday 13 January 2026  00:47:46 +0000 (0:00:00.552)       0:03:59.088 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nTuesday 13 January 2026  00:47:47 +0000 (0:00:00.698)       0:03:59.786 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nTuesday 13 January 2026  00:47:49 +0000 (0:00:01.347)       0:04:01.134 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nTuesday 13 January 2026  00:47:49 +0000 (0:00:00.646)       0:04:01.780 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nTuesday 13 January 2026  00:47:52 +0000 (0:00:03.055)       0:04:04.836 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nTuesday 13 January 2026  00:47:54 +0000 (0:00:01.562)       0:04:06.398 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nTuesday 13 January 2026  00:47:54 +0000 (0:00:00.526)       0:04:06.925 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nTuesday 13 January 2026  00:47:56 +0000 (0:00:01.471)       0:04:08.396 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nTuesday 13 January 2026  00:48:00 +0000 (0:00:03.806)       0:04:12.202 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nTuesday 13 January 2026  00:48:01 +0000 (0:00:01.256)       0:04:13.459 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nTuesday 13 January 2026  00:48:01 +0000 (0:00:00.304)       0:04:13.763 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nTuesday 13 January 2026  00:48:02 +0000 (0:00:01.005)       0:04:14.769 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nTuesday 13 January 2026  00:48:03 +0000 (0:00:00.537)       0:04:15.306 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nTuesday 13 January 2026  00:48:03 +0000 (0:00:00.317)       0:04:15.623 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nTuesday 13 January 2026  00:48:04 +0000 (0:00:00.717)       0:04:16.341 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nTuesday 13 January 2026  00:48:05 +0000 (0:00:01.548)       0:04:17.890 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nTuesday 13 January 2026  00:48:06 +0000 (0:00:01.196)       0:04:19.086 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nTuesday 13 January 2026  00:48:08 +0000 (0:00:01.952)       0:04:21.038 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nTuesday 13 January 2026  00:48:11 +0000 (0:00:02.234)       0:04:23.273 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nTuesday 13 January 2026  00:48:11 +0000 (0:00:00.738)       0:04:24.012 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nTuesday 13 January 2026  00:48:33 +0000 (0:00:21.752)       0:04:45.764 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nTuesday 13 January 2026  00:48:44 +0000 (0:00:11.068)       0:04:56.832 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nTuesday 13 January 2026  00:48:45 +0000 (0:00:00.571)       0:04:57.404 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}]) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:49:00 +0000 (0:00:15.229)       0:05:12.634 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nTuesday 13 January 2026  00:49:00 +0000 (0:00:00.313)       0:05:12.947 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nTuesday 13 January 2026  00:49:01 +0000 (0:00:00.742)       0:05:13.690 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nTuesday 13 January 2026  00:49:01 +0000 (0:00:00.319)       0:05:14.009 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nTuesday 13 January 2026  00:49:02 +0000 (0:00:00.395)       0:05:14.405 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nTuesday 13 January 2026  00:49:03 +0000 (0:00:01.238)       0:05:15.644 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mgr] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:49:04 +0000 (0:00:00.546)       0:05:16.190 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:49:04 +0000 (0:00:00.480)       0:05:16.671 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:49:05 +0000 (0:00:00.752)       0:05:17.423 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:49:06 +0000 (0:00:00.733)       0:05:18.157 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:49:06 +0000 (0:00:00.305)       0:05:18.463 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:49:06 +0000 (0:00:00.527)       0:05:18.991 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:49:07 +0000 (0:00:00.318)       0:05:19.309 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:49:08 +0000 (0:00:00.788)       0:05:20.097 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:49:08 +0000 (0:00:00.329)       0:05:20.427 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:49:08 +0000 (0:00:00.545)       0:05:20.973 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:49:09 +0000 (0:00:00.716)       0:05:21.689 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:49:10 +0000 (0:00:00.708)       0:05:22.397 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:49:10 +0000 (0:00:00.225)       0:05:22.623 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:49:10 +0000 (0:00:00.424)       0:05:23.048 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:49:11 +0000 (0:00:00.254)       0:05:23.303 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:49:11 +0000 (0:00:00.261)       0:05:23.565 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:49:11 +0000 (0:00:00.261)       0:05:23.827 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:49:12 +0000 (0:00:00.310)       0:05:24.137 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:49:12 +0000 (0:00:00.672)       0:05:24.810 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:49:13 +0000 (0:00:00.321)       0:05:25.131 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:49:13 +0000 (0:00:00.307)       0:05:25.439 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nTuesday 13 January 2026  00:49:14 +0000 (0:00:00.742)       0:05:26.181 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nTuesday 13 January 2026  00:49:14 +0000 (0:00:00.584)       0:05:26.765 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nTuesday 13 January 2026  00:49:15 +0000 (0:00:00.498)       0:05:27.263 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nTuesday 13 January 2026  00:49:15 +0000 (0:00:00.677)       0:05:27.941 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nTuesday 13 January 2026  00:49:16 +0000 (0:00:00.515)       0:05:28.456 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nTuesday 13 January 2026  00:49:27 +0000 (0:00:10.759)       0:05:39.216 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nTuesday 13 January 2026  00:49:27 +0000 (0:00:00.361)       0:05:39.578 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nTuesday 13 January 2026  00:49:29 +0000 (0:00:02.190)       0:05:41.768 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nTuesday 13 January 2026  00:49:30 +0000 (0:00:01.278)       0:05:43.047 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nTuesday 13 January 2026  00:49:31 +0000 (0:00:01.039)       0:05:44.087 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nTuesday 13 January 2026  00:49:32 +0000 (0:00:00.322)       0:05:44.409 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nTuesday 13 January 2026  00:49:32 +0000 (0:00:00.329)       0:05:44.739 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nTuesday 13 January 2026  00:49:33 +0000 (0:00:00.815)       0:05:45.554 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nTuesday 13 January 2026  00:49:33 +0000 (0:00:00.322)       0:05:45.876 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nTuesday 13 January 2026  00:49:34 +0000 (0:00:00.320)       0:05:46.197 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nTuesday 13 January 2026  00:49:34 +0000 (0:00:00.724)       0:05:46.922 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nTuesday 13 January 2026  00:49:36 +0000 (0:00:01.235)       0:05:48.158 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nTuesday 13 January 2026  00:49:37 +0000 (0:00:01.214)       0:05:49.373 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nTuesday 13 January 2026  00:49:39 +0000 (0:00:01.815)       0:05:51.188 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nTuesday 13 January 2026  00:49:41 +0000 (0:00:02.155)       0:05:53.343 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nTuesday 13 January 2026  00:49:41 +0000 (0:00:00.410)       0:05:53.754 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nTuesday 13 January 2026  00:50:17 +0000 (0:00:35.440)       0:06:29.195 ******* \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nTuesday 13 January 2026  00:50:18 +0000 (0:00:01.200)       0:06:30.396 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nTuesday 13 January 2026  00:50:18 +0000 (0:00:00.312)       0:06:30.708 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nTuesday 13 January 2026  00:50:18 +0000 (0:00:00.141)       0:06:30.850 ******* \n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nTuesday 13 January 2026  00:50:25 +0000 (0:00:06.291)       0:06:37.141 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:50:30 +0000 (0:00:05.428)       0:06:42.569 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nTuesday 13 January 2026  00:50:31 +0000 (0:00:00.808)       0:06:43.378 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nTuesday 13 January 2026  00:50:32 +0000 (0:00:00.842)       0:06:44.220 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nTuesday 13 January 2026  00:50:32 +0000 (0:00:00.348)       0:06:44.568 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nTuesday 13 January 2026  00:50:33 +0000 (0:00:01.090)       0:06:45.659 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nTuesday 13 January 2026  00:50:34 +0000 (0:00:00.600)       0:06:46.259 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-osd] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:50:34 +0000 (0:00:00.815)       0:06:47.075 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:50:35 +0000 (0:00:00.501)       0:06:47.576 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:50:36 +0000 (0:00:00.804)       0:06:48.380 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:50:36 +0000 (0:00:00.346)       0:06:48.726 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:50:37 +0000 (0:00:00.712)       0:06:49.439 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:50:38 +0000 (0:00:00.825)       0:06:50.265 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:50:39 +0000 (0:00:01.482)       0:06:51.747 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:50:39 +0000 (0:00:00.314)       0:06:52.062 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:50:40 +0000 (0:00:00.333)       0:06:52.395 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:50:40 +0000 (0:00:00.325)       0:06:52.721 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:50:41 +0000 (0:00:00.996)       0:06:53.717 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:50:42 +0000 (0:00:00.782)       0:06:54.500 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:50:42 +0000 (0:00:00.319)       0:06:54.820 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:50:43 +0000 (0:00:00.302)       0:06:55.122 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:50:43 +0000 (0:00:00.592)       0:06:55.715 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:50:43 +0000 (0:00:00.346)       0:06:56.061 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:50:44 +0000 (0:00:00.348)       0:06:56.410 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:50:44 +0000 (0:00:00.300)       0:06:56.710 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:50:45 +0000 (0:00:00.537)       0:06:57.248 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:50:45 +0000 (0:00:00.307)       0:06:57.555 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:50:45 +0000 (0:00:00.352)       0:06:57.908 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nTuesday 13 January 2026  00:50:46 +0000 (0:00:00.770)       0:06:58.678 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nTuesday 13 January 2026  00:50:46 +0000 (0:00:00.315)       0:06:58.994 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nTuesday 13 January 2026  00:50:47 +0000 (0:00:00.607)       0:06:59.601 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nTuesday 13 January 2026  00:50:48 +0000 (0:00:00.520)       0:07:00.121 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nTuesday 13 January 2026  00:50:48 +0000 (0:00:00.608)       0:07:00.730 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nTuesday 13 January 2026  00:50:48 +0000 (0:00:00.305)       0:07:01.036 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nTuesday 13 January 2026  00:50:49 +0000 (0:00:00.657)       0:07:01.693 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nTuesday 13 January 2026  00:50:49 +0000 (0:00:00.354)       0:07:02.048 ******* \n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nTuesday 13 January 2026  00:50:55 +0000 (0:00:05.525)       0:07:07.574 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nTuesday 13 January 2026  00:50:55 +0000 (0:00:00.316)       0:07:07.890 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nTuesday 13 January 2026  00:50:56 +0000 (0:00:00.524)       0:07:08.414 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nTuesday 13 January 2026  00:50:57 +0000 (0:00:01.266)       0:07:09.681 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nTuesday 13 January 2026  00:50:59 +0000 (0:00:02.105)       0:07:11.787 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nTuesday 13 January 2026  00:51:01 +0000 (0:00:01.328)       0:07:13.116 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nTuesday 13 January 2026  00:51:03 +0000 (0:00:02.308)       0:07:15.425 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Use ceph-volume to create osds] *******************************\nTuesday 13 January 2026  00:51:04 +0000 (0:00:00.826)       0:07:16.251 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nTuesday 13 January 2026  00:51:46 +0000 (0:00:42.301)       0:07:58.553 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nTuesday 13 January 2026  00:51:46 +0000 (0:00:00.323)       0:07:58.876 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nTuesday 13 January 2026  00:51:47 +0000 (0:00:00.752)       0:07:59.629 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nTuesday 13 January 2026  00:51:48 +0000 (0:00:00.721)       0:08:00.351 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nTuesday 13 January 2026  00:51:51 +0000 (0:00:03.093)       0:08:03.444 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nTuesday 13 January 2026  00:51:52 +0000 (0:00:00.746)       0:08:04.191 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nTuesday 13 January 2026  00:51:53 +0000 (0:00:01.114)       0:08:05.306 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nTuesday 13 January 2026  00:51:54 +0000 (0:00:01.169)       0:08:06.475 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nTuesday 13 January 2026  00:51:56 +0000 (0:00:01.669)       0:08:08.144 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nTuesday 13 January 2026  00:51:56 +0000 (0:00:00.597)       0:08:08.742 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nTuesday 13 January 2026  00:51:56 +0000 (0:00:00.317)       0:08:09.059 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=3)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nTuesday 13 January 2026  00:51:58 +0000 (0:00:01.078)       0:08:10.138 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nTuesday 13 January 2026  00:52:00 +0000 (0:00:02.017)       0:08:12.155 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nTuesday 13 January 2026  00:52:03 +0000 (0:00:03.878)       0:08:16.034 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nTuesday 13 January 2026  00:52:07 +0000 (0:00:03.097)       0:08:19.132 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nTuesday 13 January 2026  00:52:19 +0000 (0:00:12.335)       0:08:31.467 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:52:20 +0000 (0:00:01.044)       0:08:32.512 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nTuesday 13 January 2026  00:52:20 +0000 (0:00:00.347)       0:08:32.859 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nTuesday 13 January 2026  00:52:21 +0000 (0:00:00.518)       0:08:33.378 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nTuesday 13 January 2026  00:52:22 +0000 (0:00:00.944)       0:08:34.322 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nTuesday 13 January 2026  00:52:22 +0000 (0:00:00.323)       0:08:34.646 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nTuesday 13 January 2026  00:52:22 +0000 (0:00:00.243)       0:08:34.889 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:00.307)       0:08:35.197 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:00.245)       0:08:35.442 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:00.225)       0:08:35.668 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:00.116)       0:08:35.785 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nTuesday 13 January 2026  00:52:23 +0000 (0:00:00.199)       0:08:35.984 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nTuesday 13 January 2026  00:52:24 +0000 (0:00:00.856)       0:08:36.840 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nTuesday 13 January 2026  00:52:25 +0000 (0:00:00.412)       0:08:37.252 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nTuesday 13 January 2026  00:52:25 +0000 (0:00:00.356)       0:08:37.609 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nTuesday 13 January 2026  00:52:25 +0000 (0:00:00.214)       0:08:37.823 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role ceph-crash] ***************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:52:26 +0000 (0:00:00.966)       0:08:38.789 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:52:27 +0000 (0:00:01.276)       0:08:40.066 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:52:29 +0000 (0:00:01.033)       0:08:41.100 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:52:30 +0000 (0:00:01.257)       0:08:42.358 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:52:31 +0000 (0:00:00.833)       0:08:43.191 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:52:32 +0000 (0:00:01.063)       0:08:44.255 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:52:33 +0000 (0:00:00.868)       0:08:45.124 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:52:34 +0000 (0:00:01.464)       0:08:46.589 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:52:35 +0000 (0:00:00.589)       0:08:47.179 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:52:36 +0000 (0:00:00.993)       0:08:48.173 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:52:37 +0000 (0:00:01.163)       0:08:49.336 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:52:38 +0000 (0:00:01.219)       0:08:50.556 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:52:39 +0000 (0:00:00.584)       0:08:51.140 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:52:39 +0000 (0:00:00.700)       0:08:51.841 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:52:40 +0000 (0:00:00.519)       0:08:52.360 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:52:41 +0000 (0:00:00.735)       0:08:53.096 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:52:41 +0000 (0:00:00.537)       0:08:53.634 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:52:42 +0000 (0:00:00.727)       0:08:54.361 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:52:42 +0000 (0:00:00.509)       0:08:54.871 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:52:43 +0000 (0:00:00.824)       0:08:55.695 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:52:44 +0000 (0:00:00.476)       0:08:56.172 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nTuesday 13 January 2026  00:52:45 +0000 (0:00:01.063)       0:08:57.236 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nTuesday 13 January 2026  00:52:49 +0000 (0:00:04.212)       0:09:01.448 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nTuesday 13 January 2026  00:52:51 +0000 (0:00:02.054)       0:09:03.503 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nTuesday 13 January 2026  00:52:53 +0000 (0:00:01.895)       0:09:05.398 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nTuesday 13 January 2026  00:52:54 +0000 (0:00:01.138)       0:09:06.537 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nTuesday 13 January 2026  00:52:55 +0000 (0:00:01.049)       0:09:07.587 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nTuesday 13 January 2026  00:52:57 +0000 (0:00:01.816)       0:09:09.403 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************\nTuesday 13 January 2026  00:53:00 +0000 (0:00:03.410)       0:09:12.814 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******\nTuesday 13 January 2026  00:53:02 +0000 (0:00:01.388)       0:09:14.202 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************\nTuesday 13 January 2026  00:53:02 +0000 (0:00:00.743)       0:09:14.946 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******\nTuesday 13 January 2026  00:53:05 +0000 (0:00:02.606)       0:09:17.552 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mds] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:53:06 +0000 (0:00:00.924)       0:09:18.477 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:53:06 +0000 (0:00:00.441)       0:09:18.919 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:53:07 +0000 (0:00:00.665)       0:09:19.585 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:53:07 +0000 (0:00:00.298)       0:09:19.884 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:53:08 +0000 (0:00:00.662)       0:09:20.546 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:53:09 +0000 (0:00:00.998)       0:09:21.545 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:53:10 +0000 (0:00:00.718)       0:09:22.263 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:53:10 +0000 (0:00:00.333)       0:09:22.596 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:53:10 +0000 (0:00:00.321)       0:09:22.918 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:53:11 +0000 (0:00:00.580)       0:09:23.499 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:53:12 +0000 (0:00:00.704)       0:09:24.204 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:53:12 +0000 (0:00:00.786)       0:09:24.990 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:53:13 +0000 (0:00:00.318)       0:09:25.309 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:53:13 +0000 (0:00:00.574)       0:09:25.883 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:53:14 +0000 (0:00:00.342)       0:09:26.225 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:53:14 +0000 (0:00:00.338)       0:09:26.564 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:53:14 +0000 (0:00:00.335)       0:09:26.900 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:53:15 +0000 (0:00:00.667)       0:09:27.568 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:53:15 +0000 (0:00:00.317)       0:09:27.885 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:53:16 +0000 (0:00:00.333)       0:09:28.218 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:53:16 +0000 (0:00:00.317)       0:09:28.536 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nTuesday 13 January 2026  00:53:17 +0000 (0:00:00.823)       0:09:29.360 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule details] *********************\nTuesday 13 January 2026  00:53:17 +0000 (0:00:00.386)       0:09:29.746 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule name] ************************\nTuesday 13 January 2026  00:53:19 +0000 (0:00:02.127)       0:09:31.874 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create filesystem pools] **************************************\nTuesday 13 January 2026  00:53:20 +0000 (0:00:00.223)       0:09:32.097 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [ceph-mds : Create ceph filesystem] ***************************************\nTuesday 13 January 2026  00:53:28 +0000 (0:00:08.560)       0:09:40.657 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nTuesday 13 January 2026  00:53:32 +0000 (0:00:03.786)       0:09:44.444 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nTuesday 13 January 2026  00:53:32 +0000 (0:00:00.581)       0:09:45.026 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nTuesday 13 January 2026  00:53:33 +0000 (0:00:01.009)       0:09:46.036 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nTuesday 13 January 2026  00:53:36 +0000 (0:00:02.379)       0:09:48.415 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nTuesday 13 January 2026  00:53:37 +0000 (0:00:01.588)       0:09:50.004 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nTuesday 13 January 2026  00:53:41 +0000 (0:00:03.348)       0:09:53.352 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nTuesday 13 January 2026  00:53:41 +0000 (0:00:00.306)       0:09:53.659 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nTuesday 13 January 2026  00:53:42 +0000 (0:00:00.814)       0:09:54.474 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nTuesday 13 January 2026  00:53:42 +0000 (0:00:00.530)       0:09:55.004 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nTuesday 13 January 2026  00:53:44 +0000 (0:00:01.189)       0:09:56.194 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nTuesday 13 January 2026  00:53:45 +0000 (0:00:01.385)       0:09:57.579 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nTuesday 13 January 2026  00:53:47 +0000 (0:00:01.822)       0:09:59.401 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nTuesday 13 January 2026  00:53:49 +0000 (0:00:01.998)       0:10:01.400 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:53:50 +0000 (0:00:01.684)       0:10:03.085 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nTuesday 13 January 2026  00:53:51 +0000 (0:00:00.711)       0:10:03.796 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nTuesday 13 January 2026  00:53:52 +0000 (0:00:00.861)       0:10:04.658 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nTuesday 13 January 2026  00:53:52 +0000 (0:00:00.342)       0:10:05.000 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nTuesday 13 January 2026  00:53:54 +0000 (0:00:01.186)       0:10:06.186 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nTuesday 13 January 2026  00:53:55 +0000 (0:00:00.984)       0:10:07.171 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 13 January 2026  00:53:55 +0000 (0:00:00.878)       0:10:08.049 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 13 January 2026  00:53:56 +0000 (0:00:00.500)       0:10:08.550 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 13 January 2026  00:53:57 +0000 (0:00:00.873)       0:10:09.425 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 13 January 2026  00:53:57 +0000 (0:00:00.310)       0:10:09.736 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 13 January 2026  00:53:58 +0000 (0:00:00.680)       0:10:10.417 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 13 January 2026  00:53:59 +0000 (0:00:01.043)       0:10:11.461 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 13 January 2026  00:54:00 +0000 (0:00:00.739)       0:10:12.200 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 13 January 2026  00:54:00 +0000 (0:00:00.312)       0:10:12.512 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 13 January 2026  00:54:00 +0000 (0:00:00.334)       0:10:12.847 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 13 January 2026  00:54:01 +0000 (0:00:00.583)       0:10:13.430 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 13 January 2026  00:54:02 +0000 (0:00:00.754)       0:10:14.185 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 13 January 2026  00:54:02 +0000 (0:00:00.883)       0:10:15.069 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 13 January 2026  00:54:03 +0000 (0:00:00.337)       0:10:15.406 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 13 January 2026  00:54:03 +0000 (0:00:00.676)       0:10:16.082 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 13 January 2026  00:54:04 +0000 (0:00:00.379)       0:10:16.462 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 13 January 2026  00:54:04 +0000 (0:00:00.357)       0:10:16.820 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 13 January 2026  00:54:05 +0000 (0:00:00.320)       0:10:17.140 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 13 January 2026  00:54:05 +0000 (0:00:00.631)       0:10:17.772 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 13 January 2026  00:54:05 +0000 (0:00:00.303)       0:10:18.075 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 13 January 2026  00:54:06 +0000 (0:00:00.320)       0:10:18.396 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 13 January 2026  00:54:06 +0000 (0:00:00.320)       0:10:18.716 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nTuesday 13 January 2026  00:54:07 +0000 (0:00:00.808)       0:10:19.524 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nTuesday 13 January 2026  00:54:07 +0000 (0:00:00.520)       0:10:20.045 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nTuesday 13 January 2026  00:54:10 +0000 (0:00:02.284)       0:10:22.329 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nTuesday 13 January 2026  00:54:11 +0000 (0:00:01.522)       0:10:23.852 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nTuesday 13 January 2026  00:54:12 +0000 (0:00:00.334)       0:10:24.186 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nTuesday 13 January 2026  00:54:12 +0000 (0:00:00.545)       0:10:24.732 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nTuesday 13 January 2026  00:54:14 +0000 (0:00:01.429)       0:10:26.161 ******* \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nTuesday 13 January 2026  00:54:17 +0000 (0:00:03.924)       0:10:30.085 ******* \n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nTuesday 13 January 2026  00:54:20 +0000 (0:00:02.193)       0:10:32.279 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nTuesday 13 January 2026  00:54:21 +0000 (0:00:01.082)       0:10:33.361 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nTuesday 13 January 2026  00:54:21 +0000 (0:00:00.229)       0:10:33.591 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nTuesday 13 January 2026  00:54:22 +0000 (0:00:01.148)       0:10:34.739 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nTuesday 13 January 2026  00:54:23 +0000 (0:00:00.594)       0:10:35.334 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nTuesday 13 January 2026  00:54:53 +0000 (0:00:30.369)       0:11:05.703 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nTuesday 13 January 2026  00:54:53 +0000 (0:00:00.327)       0:11:06.030 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nTuesday 13 January 2026  00:54:54 +0000 (0:00:00.295)       0:11:06.326 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nTuesday 13 January 2026  00:54:55 +0000 (0:00:00.806)       0:11:07.132 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nTuesday 13 January 2026  00:54:55 +0000 (0:00:00.559)       0:11:07.692 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nTuesday 13 January 2026  00:54:56 +0000 (0:00:01.367)       0:11:09.060 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nTuesday 13 January 2026  00:54:58 +0000 (0:00:01.575)       0:11:10.636 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nTuesday 13 January 2026  00:55:00 +0000 (0:00:02.274)       0:11:12.910 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 13 January 2026  00:55:03 +0000 (0:00:02.951)       0:11:15.862 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nTuesday 13 January 2026  00:55:04 +0000 (0:00:00.524)       0:11:16.386 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nTuesday 13 January 2026  00:55:04 +0000 (0:00:00.514)       0:11:16.901 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nTuesday 13 January 2026  00:55:05 +0000 (0:00:00.607)       0:11:17.508 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nTuesday 13 January 2026  00:55:05 +0000 (0:00:00.333)       0:11:17.841 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nTuesday 13 January 2026  00:55:06 +0000 (0:00:00.591)       0:11:18.433 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=34  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:55:06 +0000 (0:00:00.257)       0:11:18.690 ******* \n=============================================================================== \nceph-container-common : Pulling Ceph container image ------------------- 60.89s\nceph-osd : Use ceph-volume to create osds ------------------------------ 42.30s\nceph-mgr : Wait for all mgr to be up ----------------------------------- 35.44s\nceph-rgw : Create rgw pools -------------------------------------------- 30.37s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.75s\nceph-mon : Set cluster configs ----------------------------------------- 15.23s\nceph-osd : Wait for all osd to be up ----------------------------------- 12.34s\nceph-mon : Fetch ceph initial keys ------------------------------------- 11.07s\nceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.76s\nceph-mds : Create filesystem pools -------------------------------------- 8.56s\nceph-config : Create ceph initial directories --------------------------- 7.43s\nceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.29s\nceph-osd : Apply operating system tuning -------------------------------- 5.53s\nceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.43s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 5.02s\nceph-crash : Create client.crash keyring -------------------------------- 4.21s\nceph-rgw : Create rgw keyrings ------------------------------------------ 3.92s\nceph-osd : Systemd start osd -------------------------------------------- 3.88s\nceph-mon : Generate initial monmap -------------------------------------- 3.81s\nceph-mds : Create ceph filesystem --------------------------------------- 3.79s\n\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Set kolla_action_mariadb] ************************************************\n\nTASK [Inform the user about the following task] ********************************\nTuesday 13 January 2026  00:52:29 +0000 (0:00:00.092)       0:00:00.092 ******* \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check MariaDB service] ***************************************************\nTuesday 13 January 2026  00:52:29 +0000 (0:00:00.051)       0:00:00.143 ******* \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********\nTuesday 13 January 2026  00:52:32 +0000 (0:00:03.026)       0:00:03.170 ******* \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_mariadb = kolla_action_ng] ******************************\nTuesday 13 January 2026  00:52:32 +0000 (0:00:00.082)       0:00:03.252 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:52:33 +0000 (0:00:00.166)       0:00:03.419 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:52:33 +0000 (0:00:00.292)       0:00:03.712 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nTuesday 13 January 2026  00:52:33 +0000 (0:00:00.661)       0:00:04.374 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 13 January 2026  00:52:34 +0000 (0:00:00.388)       0:00:04.762 ******* \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nTuesday 13 January 2026  00:52:35 +0000 (0:00:00.641)       0:00:05.403 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nTuesday 13 January 2026  00:52:38 +0000 (0:00:03.419)       0:00:08.823 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nTuesday 13 January 2026  00:52:39 +0000 (0:00:00.687)       0:00:09.510 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nTuesday 13 January 2026  00:52:40 +0000 (0:00:01.233)       0:00:10.744 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nTuesday 13 January 2026  00:52:44 +0000 (0:00:04.102)       0:00:14.846 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nTuesday 13 January 2026  00:52:45 +0000 (0:00:01.025)       0:00:15.872 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 13 January 2026  00:52:48 +0000 (0:00:03.472)       0:00:19.345 ******* \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nTuesday 13 January 2026  00:52:49 +0000 (0:00:00.474)       0:00:19.819 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:52:52 +0000 (0:00:03.263)       0:00:23.082 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  00:52:55 +0000 (0:00:02.926)       0:00:26.009 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Check mariadb containers] **************************************\nTuesday 13 January 2026  00:52:58 +0000 (0:00:03.340)       0:00:29.350 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nTuesday 13 January 2026  00:53:02 +0000 (0:00:03.952)       0:00:33.302 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nTuesday 13 January 2026  00:53:03 +0000 (0:00:00.794)       0:00:34.097 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nTuesday 13 January 2026  00:53:04 +0000 (0:00:00.443)       0:00:34.540 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nTuesday 13 January 2026  00:53:04 +0000 (0:00:00.284)       0:00:34.825 ******* \n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nTuesday 13 January 2026  00:53:15 +0000 (0:00:10.939)       0:00:45.765 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nTuesday 13 January 2026  00:53:15 +0000 (0:00:00.406)       0:00:46.171 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nTuesday 13 January 2026  00:53:16 +0000 (0:00:00.641)       0:00:46.812 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nTuesday 13 January 2026  00:53:16 +0000 (0:00:00.422)       0:00:47.235 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nTuesday 13 January 2026  00:53:17 +0000 (0:00:00.407)       0:00:47.642 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nTuesday 13 January 2026  00:53:17 +0000 (0:00:00.396)       0:00:48.038 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 13 January 2026  00:53:18 +0000 (0:00:00.642)       0:00:48.680 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [mariadb : Running MariaDB bootstrap container] ***************************\nTuesday 13 January 2026  00:53:18 +0000 (0:00:00.410)       0:00:49.091 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Store bootstrap host name into facts] **************************\nTuesday 13 January 2026  00:53:29 +0000 (0:00:10.549)       0:00:59.640 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 13 January 2026  00:53:29 +0000 (0:00:00.130)       0:00:59.771 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Starting first MariaDB container] *******************\nTuesday 13 January 2026  00:53:30 +0000 (0:00:00.989)       0:01:00.760 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******\nTuesday 13 January 2026  00:53:38 +0000 (0:00:07.915)       0:01:08.676 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******\nTuesday 13 January 2026  00:53:39 +0000 (0:00:01.577)       0:01:10.253 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***\nTuesday 13 January 2026  00:53:42 +0000 (0:00:02.502)       0:01:12.756 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nTuesday 13 January 2026  00:53:42 +0000 (0:00:00.117)       0:01:12.874 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nTuesday 13 January 2026  00:53:42 +0000 (0:00:00.351)       0:01:13.225 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 13 January 2026  00:53:43 +0000 (0:00:00.599)       0:01:13.825 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 13 January 2026  00:54:01 +0000 (0:00:18.426)       0:01:32.252 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 13 January 2026  00:54:17 +0000 (0:00:15.633)       0:01:47.885 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 13 January 2026  00:54:19 +0000 (0:00:02.227)       0:01:50.113 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 13 January 2026  00:54:37 +0000 (0:00:17.340)       0:02:07.453 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 13 January 2026  00:54:52 +0000 (0:00:15.602)       0:02:23.056 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 13 January 2026  00:54:55 +0000 (0:00:02.492)       0:02:25.548 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 13 January 2026  00:55:12 +0000 (0:00:17.259)       0:02:42.807 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 13 January 2026  00:55:13 +0000 (0:00:00.653)       0:02:43.461 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nTuesday 13 January 2026  00:55:15 +0000 (0:00:02.666)       0:02:46.128 ******* \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nTuesday 13 January 2026  00:55:16 +0000 (0:00:00.555)       0:02:46.684 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nTuesday 13 January 2026  00:55:18 +0000 (0:00:02.256)       0:02:48.940 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nTuesday 13 January 2026  00:55:20 +0000 (0:00:02.080)       0:02:51.021 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nTuesday 13 January 2026  00:55:23 +0000 (0:00:02.638)       0:02:53.660 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nTuesday 13 January 2026  00:55:25 +0000 (0:00:02.407)       0:02:56.067 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nTuesday 13 January 2026  00:55:29 +0000 (0:00:03.538)       0:02:59.605 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:55:29 +0000 (0:00:00.232)       0:02:59.837 ******* \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 35.77s\nmariadb : Wait for MariaDB service port liveness ----------------------- 31.24s\nmariadb : Restart MariaDB container ------------------------------------ 17.26s\nmariadb : Check MariaDB service port liveness -------------------------- 10.94s\nmariadb : Running MariaDB bootstrap container -------------------------- 10.55s\nmariadb : Starting first MariaDB container ------------------------------ 7.92s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 4.72s\nmariadb : Copying over config.json files for services ------------------- 4.10s\nmariadb : Check mariadb containers -------------------------------------- 3.95s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 3.54s\nmariadb : Copying over galera.cnf --------------------------------------- 3.47s\nmariadb : Ensuring config directories exist ----------------------------- 3.42s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 3.34s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 3.26s\nCheck MariaDB service --------------------------------------------------- 3.03s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.93s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.67s\nmariadb : Creating database backup user and setting permissions --------- 2.64s\nmariadb : Wait for first MariaDB service to sync WSREP ------------------ 2.50s\nmariadb : Granting permissions on Mariabackup database to backup user --- 2.41s\n\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.315)       0:00:00.315 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.262)       0:00:00.578 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m\n\nPLAY [Apply role horizon] ******************************************************\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.376)       0:00:00.954 ******* \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Ensuring config directories exist] *****************************\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.467)       0:00:01.422 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : Set empty custom policy] ***************************************\nTuesday 13 January 2026  00:55:36 +0000 (0:00:01.125)       0:00:02.547 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.371)       0:00:02.918 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.613)       0:00:03.531 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.263)       0:00:03.795 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.117)       0:00:03.913 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.425)       0:00:04.339 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.256)       0:00:04.596 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.120)       0:00:04.717 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:00.272)       0:00:04.989 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:00.298)       0:00:05.287 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:00.300)       0:00:05.588 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:00.259)       0:00:05.847 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:40 +0000 (0:00:00.284)       0:00:06.132 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:40 +0000 (0:00:00.106)       0:00:06.238 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:40 +0000 (0:00:00.248)       0:00:06.487 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.405)       0:00:06.892 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.130)       0:00:07.022 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.252)       0:00:07.275 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.253)       0:00:07.528 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.107)       0:00:07.636 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:41 +0000 (0:00:00.234)       0:00:07.870 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:42 +0000 (0:00:00.442)       0:00:08.312 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:42 +0000 (0:00:00.107)       0:00:08.419 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:42 +0000 (0:00:00.240)       0:00:08.660 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.258)       0:00:08.919 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.109)       0:00:09.028 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.370)       0:00:09.399 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.277)       0:00:09.676 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.104)       0:00:09.781 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 13 January 2026  00:55:44 +0000 (0:00:00.239)       0:00:10.020 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:55:44 +0000 (0:00:00.286)       0:00:10.307 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 13 January 2026  00:55:44 +0000 (0:00:00.128)       0:00:10.435 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over config.json files for services] *******************\nTuesday 13 January 2026  00:55:45 +0000 (0:00:00.520)       0:00:10.956 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [horizon : Copying over horizon.conf] *************************************\nTuesday 13 January 2026  00:55:46 +0000 (0:00:01.692)       0:00:12.648 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\nTASK [horizon : Copying over kolla-settings.py] ********************************\nTuesday 13 January 2026  00:55:48 +0000 (0:00:01.716)       0:00:14.365 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over custom-settings.py] *******************************\nTuesday 13 January 2026  00:55:50 +0000 (0:00:02.318)       0:00:16.683 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over existing policy file] *****************************\nTuesday 13 January 2026  00:55:52 +0000 (0:00:02.063)       0:00:18.747 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over custom themes] ************************************\nTuesday 13 January 2026  00:55:53 +0000 (0:00:00.298)       0:00:19.045 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 13 January 2026  00:55:53 +0000 (0:00:00.310)       0:00:19.356 ******* \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over extra CA certificates] ********\nTuesday 13 January 2026  00:55:54 +0000 (0:00:00.844)       0:00:20.201 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:55:55 +0000 (0:00:01.462)       0:00:21.664 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  00:55:56 +0000 (0:00:00.629)       0:00:22.293 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Deploy horizon container] **************************************\nTuesday 13 January 2026  00:55:57 +0000 (0:00:00.950)       0:00:23.243 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 13 January 2026  00:55:59 +0000 (0:00:01.652)       0:00:24.896 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 13 January 2026  00:55:59 +0000 (0:00:00.302)       0:00:25.198 ******* \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Creating Horizon database] *************************************\nTuesday 13 January 2026  00:55:59 +0000 (0:00:00.516)       0:00:25.715 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Creating Horizon database user and setting permissions] ********\nTuesday 13 January 2026  00:56:02 +0000 (0:00:02.774)       0:00:28.489 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Running Horizon bootstrap container] ***************************\nTuesday 13 January 2026  00:56:05 +0000 (0:00:03.164)       0:00:31.654 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 13 January 2026  00:56:22 +0000 (0:00:16.846)       0:00:48.501 ******* \n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 13 January 2026  00:56:22 +0000 (0:00:00.075)       0:00:48.577 ******* \n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 13 January 2026  00:56:22 +0000 (0:00:00.081)       0:00:48.658 ******* \n\nRUNNING HANDLER [horizon : Restart horizon container] **************************\nTuesday 13 January 2026  00:56:22 +0000 (0:00:00.069)       0:00:48.727 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:57:16 +0000 (0:00:53.330)       0:01:42.058 ******* \n=============================================================================== \nhorizon : Restart horizon container ------------------------------------ 53.33s\nhorizon : Running Horizon bootstrap container -------------------------- 16.85s\nhorizon : Creating Horizon database user and setting permissions -------- 3.17s\nhorizon : Creating Horizon database ------------------------------------- 2.77s\nhorizon : Copying over kolla-settings.py -------------------------------- 2.32s\nhorizon : Copying over custom-settings.py ------------------------------- 2.06s\nhorizon : Copying over horizon.conf ------------------------------------- 1.72s\nhorizon : Copying over config.json files for services ------------------- 1.69s\nhorizon : Deploy horizon container -------------------------------------- 1.65s\nservice-cert-copy : horizon | Copying over extra CA certificates -------- 1.46s\nhorizon : Ensuring config directories exist ----------------------------- 1.13s\nservice-cert-copy : horizon | Copying over backend internal TLS key ----- 0.95s\nhorizon : include_tasks ------------------------------------------------- 0.84s\nservice-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.63s\nhorizon : include_tasks ------------------------------------------------- 0.61s\nhorizon : Update custom policy file name -------------------------------- 0.52s\nhorizon : include_tasks ------------------------------------------------- 0.52s\nhorizon : include_tasks ------------------------------------------------- 0.47s\nhorizon : Update policy file name --------------------------------------- 0.44s\nhorizon : Update custom policy file name -------------------------------- 0.43s\n\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Create ceph pools] *******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nTuesday 13 January 2026  00:55:11 +0000 (0:00:00.561)       0:00:00.561 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nTuesday 13 January 2026  00:55:12 +0000 (0:00:00.550)       0:00:01.111 ******* \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nTuesday 13 January 2026  00:55:12 +0000 (0:00:00.625)       0:00:01.737 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nTuesday 13 January 2026  00:55:12 +0000 (0:00:00.252)       0:00:01.989 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nTuesday 13 January 2026  00:55:13 +0000 (0:00:00.841)       0:00:02.831 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nTuesday 13 January 2026  00:55:14 +0000 (0:00:00.284)       0:00:03.116 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nTuesday 13 January 2026  00:55:14 +0000 (0:00:00.249)       0:00:03.365 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nTuesday 13 January 2026  00:55:14 +0000 (0:00:00.273)       0:00:03.639 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nTuesday 13 January 2026  00:55:15 +0000 (0:00:00.502)       0:00:04.142 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nTuesday 13 January 2026  00:55:15 +0000 (0:00:00.313)       0:00:04.456 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nTuesday 13 January 2026  00:55:16 +0000 (0:00:00.637)       0:00:05.093 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nTuesday 13 January 2026  00:55:16 +0000 (0:00:00.436)       0:00:05.529 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nTuesday 13 January 2026  00:55:18 +0000 (0:00:02.055)       0:00:07.585 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nTuesday 13 January 2026  00:55:19 +0000 (0:00:00.657)       0:00:08.242 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nTuesday 13 January 2026  00:55:20 +0000 (0:00:00.825)       0:00:09.068 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nTuesday 13 January 2026  00:55:20 +0000 (0:00:00.417)       0:00:09.486 ******* \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '000f2103ee99', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-01-13 00:55:17.152761', 'end': '2026-01-13 00:55:17.180036', 'delta': '0:00:00.027275', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['000f2103ee99'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '84ac1be37239', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-01-13 00:55:17.884463', 'end': '2026-01-13 00:55:17.909365', 'delta': '0:00:00.024902', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['84ac1be37239'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '01f2672e35a4', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-01-13 00:55:18.404904', 'end': '2026-01-13 00:55:18.432868', 'delta': '0:00:00.027964', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['01f2672e35a4'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nTuesday 13 January 2026  00:55:20 +0000 (0:00:00.198)       0:00:09.685 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nTuesday 13 January 2026  00:55:21 +0000 (0:00:00.439)       0:00:10.124 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nTuesday 13 January 2026  00:55:22 +0000 (0:00:01.873)       0:00:11.998 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nTuesday 13 January 2026  00:55:23 +0000 (0:00:00.295)       0:00:12.294 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 13 January 2026  00:55:23 +0000 (0:00:00.391)       0:00:12.685 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nTuesday 13 January 2026  00:55:24 +0000 (0:00:00.489)       0:00:13.174 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nTuesday 13 January 2026  00:55:24 +0000 (0:00:00.113)       0:00:13.288 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 13 January 2026  00:55:24 +0000 (0:00:00.233)       0:00:13.521 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nTuesday 13 January 2026  00:55:24 +0000 (0:00:00.285)       0:00:13.807 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nTuesday 13 January 2026  00:55:25 +0000 (0:00:00.310)       0:00:14.117 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nTuesday 13 January 2026  00:55:25 +0000 (0:00:00.528)       0:00:14.646 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nTuesday 13 January 2026  00:55:25 +0000 (0:00:00.311)       0:00:14.957 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nTuesday 13 January 2026  00:55:26 +0000 (0:00:00.307)       0:00:15.265 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nTuesday 13 January 2026  00:55:26 +0000 (0:00:00.322)       0:00:15.587 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nTuesday 13 January 2026  00:55:27 +0000 (0:00:00.564)       0:00:16.151 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nTuesday 13 January 2026  00:55:27 +0000 (0:00:00.720)       0:00:16.872 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nTuesday 13 January 2026  00:55:28 +0000 (0:00:00.562)       0:00:17.434 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nTuesday 13 January 2026  00:55:29 +0000 (0:00:00.656)       0:00:18.090 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 13 January 2026  00:55:29 +0000 (0:00:00.541)       0:00:18.632 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 13 January 2026  00:55:30 +0000 (0:00:00.613)       0:00:19.246 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 13 January 2026  00:55:30 +0000 (0:00:00.310)       0:00:19.556 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 13 January 2026  00:55:30 +0000 (0:00:00.429)       0:00:19.985 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nTuesday 13 January 2026  00:55:31 +0000 (0:00:00.547)       0:00:20.533 ******* \n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nTuesday 13 January 2026  00:55:32 +0000 (0:00:01.011)       0:00:21.545 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nTuesday 13 January 2026  00:55:32 +0000 (0:00:00.370)       0:00:21.915 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 13 January 2026  00:55:33 +0000 (0:00:00.799)       0:00:22.715 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 13 January 2026  00:55:33 +0000 (0:00:00.303)       0:00:23.019 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.379)       0:00:23.398 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.257)       0:00:23.655 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.495)       0:00:24.151 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.345)       0:00:24.497 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.311)       0:00:24.808 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 13 January 2026  00:55:36 +0000 (0:00:00.293)       0:00:25.101 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 13 January 2026  00:55:36 +0000 (0:00:00.248)       0:00:25.350 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nTuesday 13 January 2026  00:55:36 +0000 (0:00:00.451)       0:00:25.801 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.820)       0:00:26.622 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Include tasks from the ceph-osd role] ************************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:01.620)       0:00:28.242 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m\n\nTASK [create openstack pool(s)] ************************************************\nTuesday 13 January 2026  00:55:39 +0000 (0:00:00.376)       0:00:28.619 ******* \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [generate keys] ***********************************************************\nTuesday 13 January 2026  00:56:26 +0000 (0:00:46.576)       0:01:15.195 ******* \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [get keys from monitors] **************************************************\nTuesday 13 January 2026  00:56:49 +0000 (0:00:23.636)       0:01:38.832 ******* \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [copy ceph key(s) if needed] **********************************************\nTuesday 13 January 2026  00:57:01 +0000 (0:00:12.096)       0:01:50.929 ******* \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:57:19 +0000 (0:00:17.775)       0:02:08.704 ******* \n=============================================================================== \ncreate openstack pool(s) ----------------------------------------------- 46.58s\ngenerate keys ---------------------------------------------------------- 23.64s\ncopy ceph key(s) if needed --------------------------------------------- 17.78s\nget keys from monitors ------------------------------------------------- 12.10s\nceph-facts : Find a running mon container ------------------------------- 2.06s\nceph-facts : Get current fsid if cluster is already running ------------- 1.87s\nceph-facts : Set_fact ceph_admin_command -------------------------------- 1.62s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 1.01s\nceph-facts : Check if podman binary is present -------------------------- 0.84s\nceph-facts : Check if the ceph mon socket is in-use --------------------- 0.83s\nceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.82s\nceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.80s\nceph-facts : Collect existed devices ------------------------------------ 0.72s\nceph-facts : Check for a ceph mon socket -------------------------------- 0.66s\nceph-facts : Check if the ceph conf exists ------------------------------ 0.66s\nceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.64s\nceph-facts : Check if it is atomic host --------------------------------- 0.63s\nceph-facts : Read osd pool default crush rule --------------------------- 0.61s\nceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks --- 0.56s\nceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.56s\n\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy ceph keys to the configuration repository] **************************\n\nTASK [Check if ceph keys exist] ************************************************\nTuesday 13 January 2026  00:57:24 +0000 (0:00:00.114)       0:00:00.114 ******* \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Fetch all ceph keys] *****************************************************\nTuesday 13 January 2026  00:57:28 +0000 (0:00:04.439)       0:00:04.553 ******* \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Create share directory] **************************************************\nTuesday 13 January 2026  00:57:33 +0000 (0:00:04.412)       0:00:08.966 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Write ceph keys to the share directory] **********************************\nTuesday 13 January 2026  00:57:34 +0000 (0:00:01.028)       0:00:09.994 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Check if target directories exist] ***************************************\nTuesday 13 January 2026  00:57:47 +0000 (0:00:13.266)       0:00:23.261 ******* \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m\n\nTASK [Write ceph keys to the configuration directory] **************************\nTuesday 13 January 2026  00:57:51 +0000 (0:00:04.075)       0:00:27.336 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:57:58 +0000 (0:00:07.208)       0:00:34.545 ******* \n=============================================================================== \nWrite ceph keys to the share directory --------------------------------- 13.27s\nWrite ceph keys to the configuration directory -------------------------- 7.21s\nCheck if ceph keys exist ------------------------------------------------ 4.44s\nFetch all ceph keys ----------------------------------------------------- 4.41s\nCheck if target directories exist --------------------------------------- 4.08s\nCreate share directory -------------------------------------------------- 1.03s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.256)       0:00:00.256 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.232)       0:00:00.488 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Apply role keystone] *****************************************************\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:55:34 +0000 (0:00:00.355)       0:00:00.843 ******* \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Ensuring config directories exist] ****************************\nTuesday 13 January 2026  00:55:35 +0000 (0:00:00.457)       0:00:01.301 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Check if policies shall be overwritten] ***********************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:01.705)       0:00:03.006 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Set keystone policy file] *************************************\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.112)       0:00:03.118 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Check if Keystone domain-specific config is supplied] *********\nTuesday 13 January 2026  00:55:37 +0000 (0:00:00.390)       0:00:03.509 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.738)       0:00:04.248 ******* \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over extra CA certificates] *******\nTuesday 13 January 2026  00:55:38 +0000 (0:00:00.499)       0:00:04.747 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:55:42 +0000 (0:00:03.370)       0:00:08.118 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****\nTuesday 13 January 2026  00:55:42 +0000 (0:00:00.642)       0:00:08.760 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over config.json files for services] ******************\nTuesday 13 January 2026  00:55:43 +0000 (0:00:00.639)       0:00:09.400 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying over keystone.conf] ***********************************\nTuesday 13 January 2026  00:55:46 +0000 (0:00:03.447)       0:00:12.848 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying keystone-startup script for keystone] *****************\nTuesday 13 January 2026  00:55:52 +0000 (0:00:05.521)       0:00:18.370 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Create Keystone domain-specific config directory] *************\nTuesday 13 January 2026  00:55:53 +0000 (0:00:01.428)       0:00:19.798 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Get file list in custom domains folder] ***********************\nTuesday 13 January 2026  00:55:54 +0000 (0:00:00.599)       0:00:20.398 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying Keystone Domain specific settings] ********************\nTuesday 13 January 2026  00:55:54 +0000 (0:00:00.284)       0:00:20.682 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over existing policy file] ****************************\nTuesday 13 January 2026  00:55:55 +0000 (0:00:00.595)       0:00:21.278 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:55:55 +0000 (0:00:00.568)       0:00:21.847 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over wsgi-keystone.conf] ******************************\nTuesday 13 January 2026  00:55:56 +0000 (0:00:00.301)       0:00:22.148 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\nTASK [keystone : Checking whether keystone-paste.ini file exists] **************\nTuesday 13 January 2026  00:55:57 +0000 (0:00:01.800)       0:00:23.949 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Copying over keystone-paste.ini] ******************************\nTuesday 13 January 2026  00:55:58 +0000 (0:00:00.900)       0:00:24.849 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Generate the required cron jobs for the node] *****************\nTuesday 13 January 2026  00:55:59 +0000 (0:00:00.852)       0:00:25.701 ******* \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***\nTuesday 13 January 2026  00:56:00 +0000 (0:00:01.059)       0:00:26.761 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying files for keystone-fernet] ****************************\nTuesday 13 January 2026  00:56:01 +0000 (0:00:00.296)       0:00:27.058 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [keystone : Copying files for keystone-ssh] *******************************\nTuesday 13 January 2026  00:56:09 +0000 (0:00:08.632)       0:00:35.690 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\nTASK [keystone : Check keystone containers] ************************************\nTuesday 13 January 2026  00:56:12 +0000 (0:00:03.106)       0:00:38.797 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:56:15 +0000 (0:00:02.376)       0:00:41.173 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Creating keystone database] ***********************************\nTuesday 13 January 2026  00:56:15 +0000 (0:00:00.294)       0:00:41.467 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating Keystone database user and setting permissions] ******\nTuesday 13 January 2026  00:56:17 +0000 (0:00:02.494)       0:00:43.961 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Checking for any running keystone_fernet containers] **********\nTuesday 13 January 2026  00:56:20 +0000 (0:00:02.557)       0:00:46.519 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [keystone : Group nodes where keystone_fernet is running] *****************\nTuesday 13 January 2026  00:56:21 +0000 (0:00:01.208)       0:00:47.727 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***\nTuesday 13 January 2026  00:56:22 +0000 (0:00:00.305)       0:00:48.033 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Running Keystone bootstrap container] *************************\nTuesday 13 January 2026  00:56:22 +0000 (0:00:00.347)       0:00:48.381 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Running Keystone fernet bootstrap container] ******************\nTuesday 13 January 2026  00:56:36 +0000 (0:00:14.013)       0:01:02.394 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:56:47 +0000 (0:00:11.244)       0:01:13.638 ******* \n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:56:47 +0000 (0:00:00.065)       0:01:13.704 ******* \n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:56:47 +0000 (0:00:00.077)       0:01:13.781 ******* \n\nRUNNING HANDLER [keystone : Restart keystone-ssh container] ********************\nTuesday 13 January 2026  00:56:47 +0000 (0:00:00.066)       0:01:13.848 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone-fernet container] *****************\nTuesday 13 January 2026  00:57:03 +0000 (0:00:15.296)       0:01:29.145 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone container] ************************\nTuesday 13 January 2026  00:57:08 +0000 (0:00:05.427)       0:01:34.572 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:57:19 +0000 (0:00:10.936)       0:01:45.509 ******* \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Waiting for Keystone SSH port to be UP] ***********************\nTuesday 13 January 2026  00:57:20 +0000 (0:00:00.798)       0:01:46.307 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Run key distribution] *****************************************\nTuesday 13 January 2026  00:57:21 +0000 (0:00:00.812)       0:01:47.120 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating admin project, user, role, service, and endpoint] ****\nTuesday 13 January 2026  00:57:22 +0000 (0:00:01.571)       0:01:48.691 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating services] **********************\nTuesday 13 January 2026  00:57:34 +0000 (0:00:12.222)       0:02:00.913 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m\n\nTASK [service-ks-register : keystone | Creating endpoints] *********************\nTuesday 13 January 2026  00:58:04 +0000 (0:00:29.303)       0:02:30.217 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating projects] **********************\nTuesday 13 January 2026  00:58:10 +0000 (0:00:06.295)       0:02:36.512 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating users] *************************\nTuesday 13 January 2026  00:58:10 +0000 (0:00:00.127)       0:02:36.640 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating roles] *************************\nTuesday 13 January 2026  00:58:10 +0000 (0:00:00.109)       0:02:36.749 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Granting user roles] ********************\nTuesday 13 January 2026  00:58:10 +0000 (0:00:00.136)       0:02:36.885 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating default user role] ***********************************\nTuesday 13 January 2026  00:58:11 +0000 (0:00:00.830)       0:02:37.716 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 13 January 2026  00:58:15 +0000 (0:00:03.662)       0:02:41.378 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:58:15 +0000 (0:00:00.409)       0:02:41.788 ******* \n=============================================================================== \nservice-ks-register : keystone | Creating services --------------------- 29.30s\nkeystone : Restart keystone-ssh container ------------------------------ 15.30s\nkeystone : Running Keystone bootstrap container ------------------------ 14.01s\nkeystone : Creating admin project, user, role, service, and endpoint --- 12.22s\nkeystone : Running Keystone fernet bootstrap container ----------------- 11.24s\nkeystone : Restart keystone container ---------------------------------- 10.94s\nkeystone : Copying files for keystone-fernet ---------------------------- 8.63s\nservice-ks-register : keystone | Creating endpoints --------------------- 6.30s\nkeystone : Copying over keystone.conf ----------------------------------- 5.52s\nkeystone : Restart keystone-fernet container ---------------------------- 5.43s\nkeystone : Creating default user role ----------------------------------- 3.66s\nkeystone : Copying over config.json files for services ------------------ 3.45s\nservice-cert-copy : keystone | Copying over extra CA certificates ------- 3.37s\nkeystone : Copying files for keystone-ssh ------------------------------- 3.11s\nkeystone : Creating Keystone database user and setting permissions ------ 2.56s\nkeystone : Creating keystone database ----------------------------------- 2.49s\nkeystone : Check keystone containers ------------------------------------ 2.38s\nkeystone : Copying over wsgi-keystone.conf ------------------------------ 1.80s\nkeystone : Ensuring config directories exist ---------------------------- 1.71s\nkeystone : Run key distribution ----------------------------------------- 1.57s\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nTuesday 13 January 2026  00:58:03 +0000 (0:00:00.235)       0:00:00.235 ******* \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nTuesday 13 January 2026  00:58:03 +0000 (0:00:00.240)       0:00:00.475 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nTuesday 13 January 2026  00:58:04 +0000 (0:00:01.330)       0:00:01.806 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nTuesday 13 January 2026  00:58:06 +0000 (0:00:01.525)       0:00:03.332 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nTuesday 13 January 2026  00:58:07 +0000 (0:00:00.952)       0:00:04.284 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nTuesday 13 January 2026  00:58:08 +0000 (0:00:00.943)       0:00:05.228 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nTuesday 13 January 2026  00:58:43 +0000 (0:00:35.321)       0:00:40.549 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nTuesday 13 January 2026  00:58:47 +0000 (0:00:04.039)       0:00:44.589 ******* \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nTuesday 13 January 2026  00:58:48 +0000 (0:00:00.541)       0:00:45.130 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nTuesday 13 January 2026  00:58:48 +0000 (0:00:00.132)       0:00:45.263 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******\nTuesday 13 January 2026  00:58:48 +0000 (0:00:00.577)       0:00:45.841 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***\nTuesday 13 January 2026  00:58:51 +0000 (0:00:02.034)       0:00:47.876 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******\nTuesday 13 January 2026  00:58:51 +0000 (0:00:00.754)       0:00:48.630 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****\nTuesday 13 January 2026  00:58:52 +0000 (0:00:00.727)       0:00:49.357 ******* \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:58:54 +0000 (0:00:01.628)       0:00:50.986 ******* \n=============================================================================== \nosism.services.cephclient : Manage cephclient service ------------------ 35.32s\nosism.services.cephclient : Copy wrapper scripts ------------------------ 4.04s\nosism.services.cephclient : Restart cephclient service ------------------ 2.03s\nosism.services.cephclient : Copy bash completion scripts ---------------- 1.63s\nosism.services.cephclient : Copy configuration files -------------------- 1.53s\nosism.services.cephclient : Create required directories ----------------- 1.33s\nosism.services.cephclient : Copy keyring file --------------------------- 0.95s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 0.94s\nosism.services.cephclient : Ensure that all containers are up ----------- 0.75s\nosism.services.cephclient : Wait for an healthy service ----------------- 0.73s\nosism.services.cephclient : Include rook task --------------------------- 0.58s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.54s\nosism.services.cephclient : Include container tasks --------------------- 0.24s\nosism.services.cephclient : Include package tasks ----------------------- 0.13s\n\n\nPLAY [Download ironic ipa images] **********************************************\n\nTASK [Ensure the destination directory exists] *********************************\nTuesday 13 January 2026  00:58:21 +0000 (0:00:00.212)       0:00:00.212 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent initramfs] *****************************************\nTuesday 13 January 2026  00:58:22 +0000 (0:00:01.242)       0:00:01.455 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent kernel] ********************************************\nTuesday 13 January 2026  00:58:55 +0000 (0:00:32.735)       0:00:34.191 ******* \n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (3 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:59:22 +0000 (0:00:27.198)       0:01:01.389 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:59:23 +0000 (0:00:00.699)       0:01:02.088 ******* \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m\n\nPLAY [Apply role ironic] *******************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  00:59:24 +0000 (0:00:00.985)       0:01:03.073 ******* \n=============================================================================== \nDownload ironic-agent initramfs ---------------------------------------- 32.74s\nDownload ironic-agent kernel ------------------------------------------- 27.20s\nEnsure the destination directory exists --------------------------------- 1.24s\nGroup hosts based on enabled services ----------------------------------- 0.99s\nGroup hosts based on Kolla action --------------------------------------- 0.70s\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:58:22 +0000 (0:00:00.346)       0:00:00.346 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.445)       0:00:00.791 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m\n\nPLAY [Apply role barbican] *****************************************************\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.536)       0:00:01.328 ******* \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : barbican | Creating services] **********************\nTuesday 13 January 2026  00:58:24 +0000 (0:00:00.487)       0:00:01.816 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m\n\nTASK [service-ks-register : barbican | Creating endpoints] *********************\nTuesday 13 January 2026  00:58:28 +0000 (0:00:03.974)       0:00:05.791 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating projects] **********************\nTuesday 13 January 2026  00:58:35 +0000 (0:00:07.737)       0:00:13.528 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating users] *************************\nTuesday 13 January 2026  00:58:39 +0000 (0:00:03.390)       0:00:16.918 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating roles] *************************\nTuesday 13 January 2026  00:58:43 +0000 (0:00:04.393)       0:00:21.312 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m\n\nTASK [service-ks-register : barbican | Granting user roles] ********************\nTuesday 13 January 2026  00:59:01 +0000 (0:00:17.685)       0:00:38.997 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m\n\nTASK [barbican : Ensuring config directories exist] ****************************\nTuesday 13 January 2026  00:59:06 +0000 (0:00:04.792)       0:00:43.789 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Ensuring vassals config directories exist] ********************\nTuesday 13 January 2026  00:59:08 +0000 (0:00:02.354)       0:00:46.144 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m\n\nTASK [barbican : Check if policies shall be overwritten] ***********************\nTuesday 13 January 2026  00:59:10 +0000 (0:00:01.742)       0:00:47.886 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Set barbican policy file] *************************************\nTuesday 13 January 2026  00:59:10 +0000 (0:00:00.102)       0:00:47.988 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 13 January 2026  00:59:10 +0000 (0:00:00.376)       0:00:48.365 ******* \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over extra CA certificates] *******\nTuesday 13 January 2026  00:59:11 +0000 (0:00:00.452)       0:00:48.817 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:59:14 +0000 (0:00:03.018)       0:00:51.836 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****\nTuesday 13 January 2026  00:59:15 +0000 (0:00:01.274)       0:00:53.111 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over config.json files for services] ******************\nTuesday 13 January 2026  00:59:16 +0000 (0:00:01.434)       0:00:54.545 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over barbican-api.ini] ********************************\nTuesday 13 January 2026  00:59:21 +0000 (0:00:04.191)       0:00:58.737 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [barbican : Checking whether barbican-api-paste.ini file exists] **********\nTuesday 13 January 2026  00:59:23 +0000 (0:00:02.499)       0:01:01.236 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [barbican : Copying over barbican-api-paste.ini] **************************\nTuesday 13 January 2026  00:59:24 +0000 (0:00:01.097)       0:01:02.333 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over barbican.conf] ***********************************\nTuesday 13 January 2026  00:59:25 +0000 (0:00:00.662)       0:01:02.996 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over existing policy file] ****************************\nTuesday 13 January 2026  00:59:35 +0000 (0:00:10.273)       0:01:13.269 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Check barbican containers] ************************************\nTuesday 13 January 2026  00:59:36 +0000 (0:00:01.211)       0:01:14.481 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 13 January 2026  00:59:40 +0000 (0:00:03.821)       0:01:18.302 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Creating barbican database] ***********************************\nTuesday 13 January 2026  00:59:41 +0000 (0:00:00.474)       0:01:18.777 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Creating barbican database user and setting permissions] ******\nTuesday 13 January 2026  00:59:43 +0000 (0:00:02.494)       0:01:21.271 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Running barbican bootstrap container] *************************\nTuesday 13 January 2026  00:59:46 +0000 (0:00:02.649)       0:01:23.921 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:59:59 +0000 (0:00:13.110)       0:01:37.031 ******* \n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:59:59 +0000 (0:00:00.121)       0:01:37.152 ******* \n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 13 January 2026  00:59:59 +0000 (0:00:00.118)       0:01:37.271 ******* \n\nRUNNING HANDLER [barbican : Restart barbican-api container] ********************\nTuesday 13 January 2026  00:59:59 +0000 (0:00:00.066)       0:01:37.337 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******\nTuesday 13 January 2026  01:00:07 +0000 (0:00:07.277)       0:01:44.614 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-worker container] *****************\nTuesday 13 January 2026  01:00:12 +0000 (0:00:05.055)       0:01:49.671 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:00:20 +0000 (0:00:08.673)       0:01:58.345 ******* \n=============================================================================== \nservice-ks-register : barbican | Creating roles ------------------------ 17.69s\nbarbican : Running barbican bootstrap container ------------------------ 13.11s\nbarbican : Copying over barbican.conf ---------------------------------- 10.27s\nbarbican : Restart barbican-worker container ---------------------------- 8.67s\nservice-ks-register : barbican | Creating endpoints --------------------- 7.74s\nbarbican : Restart barbican-api container ------------------------------- 7.28s\nbarbican : Restart barbican-keystone-listener container ----------------- 5.06s\nservice-ks-register : barbican | Granting user roles -------------------- 4.79s\nservice-ks-register : barbican | Creating users ------------------------- 4.39s\nbarbican : Copying over config.json files for services ------------------ 4.19s\nservice-ks-register : barbican | Creating services ---------------------- 3.97s\nbarbican : Check barbican containers ------------------------------------ 3.82s\nservice-ks-register : barbican | Creating projects ---------------------- 3.39s\nservice-cert-copy : barbican | Copying over extra CA certificates ------- 3.02s\nbarbican : Creating barbican database user and setting permissions ------ 2.65s\nbarbican : Copying over barbican-api.ini -------------------------------- 2.50s\nbarbican : Creating barbican database ----------------------------------- 2.49s\nbarbican : Ensuring config directories exist ---------------------------- 2.35s\nbarbican : Ensuring vassals config directories exist -------------------- 1.74s\nservice-cert-copy : barbican | Copying over backend internal TLS key ---- 1.43s\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Bootstraph ceph dashboard] ***********************************************\n\nTASK [Disable the ceph dashboard] **********************************************\nTuesday 13 January 2026  00:58:58 +0000 (0:00:00.287)       0:00:00.287 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/ssl to false] ******************************************\nTuesday 13 January 2026  00:59:00 +0000 (0:00:01.892)       0:00:02.179 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_port to 7000] ***********************************\nTuesday 13 January 2026  00:59:02 +0000 (0:00:01.350)       0:00:03.530 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************\nTuesday 13 January 2026  00:59:03 +0000 (0:00:01.031)       0:00:04.562 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_behaviour to error] ****************************\nTuesday 13 January 2026  00:59:04 +0000 (0:00:01.618)       0:00:06.180 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_error_status_code to 404] **********************\nTuesday 13 January 2026  00:59:06 +0000 (0:00:01.226)       0:00:07.407 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable the ceph dashboard] ***********************************************\nTuesday 13 January 2026  00:59:07 +0000 (0:00:01.012)       0:00:08.420 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Write ceph_dashboard_password to temporary file] *************************\nTuesday 13 January 2026  00:59:08 +0000 (0:00:01.121)       0:00:09.541 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Create admin user] *******************************************************\nTuesday 13 January 2026  00:59:09 +0000 (0:00:01.194)       0:00:10.736 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Remove temporary file for ceph_dashboard_password] ***********************\nTuesday 13 January 2026  01:00:10 +0000 (0:01:00.865)       0:01:11.602 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 13 January 2026  01:00:10 +0000 (0:00:00.126)       0:01:11.728 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 13 January 2026  01:00:21 +0000 (0:00:11.353)       0:01:23.082 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 13 January 2026  01:00:32 +0000 (0:00:11.150)       0:01:34.232 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:00:34 +0000 (0:00:01.246)       0:01:35.479 ******* \n=============================================================================== \nCreate admin user ------------------------------------------------------ 60.87s\nRestart ceph manager service ------------------------------------------- 23.75s\nDisable the ceph dashboard ---------------------------------------------- 1.89s\nSet mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.62s\nSet mgr/dashboard/ssl to false ------------------------------------------ 1.35s\nSet mgr/dashboard/standby_behaviour to error ---------------------------- 1.23s\nWrite ceph_dashboard_password to temporary file ------------------------- 1.19s\nEnable the ceph dashboard ----------------------------------------------- 1.12s\nSet mgr/dashboard/server_port to 7000 ----------------------------------- 1.03s\nSet mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.01s\nRemove temporary file for ceph_dashboard_password ----------------------- 0.13s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:59:31 +0000 (0:00:00.215)       0:00:00.215 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:59:32 +0000 (0:00:00.227)       0:00:00.442 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m\n\nPLAY [Apply role placement] ****************************************************\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 13 January 2026  00:59:32 +0000 (0:00:00.371)       0:00:00.814 ******* \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : placement | Creating services] *********************\nTuesday 13 January 2026  00:59:33 +0000 (0:00:00.693)       0:00:01.507 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m\n\nTASK [service-ks-register : placement | Creating endpoints] ********************\nTuesday 13 January 2026  00:59:37 +0000 (0:00:04.225)       0:00:05.733 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m\n\nTASK [service-ks-register : placement | Creating projects] *********************\nTuesday 13 January 2026  00:59:44 +0000 (0:00:06.768)       0:00:12.502 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating users] ************************\nTuesday 13 January 2026  00:59:47 +0000 (0:00:03.506)       0:00:16.009 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating roles] ************************\nTuesday 13 January 2026  00:59:52 +0000 (0:00:04.407)       0:00:20.416 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : placement | Granting user roles] *******************\nTuesday 13 January 2026  00:59:56 +0000 (0:00:04.351)       0:00:24.768 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 13 January 2026  01:00:00 +0000 (0:00:04.456)       0:00:29.225 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Ensuring config directories exist] ***************************\nTuesday 13 January 2026  01:00:01 +0000 (0:00:00.528)       0:00:29.753 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Check if policies shall be overwritten] **********************\nTuesday 13 January 2026  01:00:02 +0000 (0:00:01.276)       0:00:31.030 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [placement : Set placement policy file] ***********************************\nTuesday 13 January 2026  01:00:02 +0000 (0:00:00.210)       0:00:31.240 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 13 January 2026  01:00:03 +0000 (0:00:00.527)       0:00:31.768 ******* \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over extra CA certificates] ******\nTuesday 13 January 2026  01:00:03 +0000 (0:00:00.471)       0:00:32.240 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:00:05 +0000 (0:00:01.683)       0:00:33.923 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS key] ***\nTuesday 13 January 2026  01:00:06 +0000 (0:00:00.937)       0:00:34.861 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over config.json files for services] *****************\nTuesday 13 January 2026  01:00:07 +0000 (0:00:01.537)       0:00:36.398 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement.conf] *********************************\nTuesday 13 January 2026  01:00:09 +0000 (0:00:01.644)       0:00:38.043 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement-api wsgi configuration] ***************\nTuesday 13 January 2026  01:00:12 +0000 (0:00:03.354)       0:00:41.397 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\nTASK [placement : Copying over migrate-db.rc.j2 configuration] *****************\nTuesday 13 January 2026  01:00:14 +0000 (0:00:01.679)       0:00:43.077 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over existing policy file] ***************************\nTuesday 13 January 2026  01:00:16 +0000 (0:00:01.726)       0:00:44.804 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Check placement containers] **********************************\nTuesday 13 January 2026  01:00:16 +0000 (0:00:00.434)       0:00:45.239 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Creating placement databases] ********************************\nTuesday 13 January 2026  01:00:17 +0000 (0:00:01.018)       0:00:46.257 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Creating placement databases user and setting permissions] ***\nTuesday 13 January 2026  01:00:20 +0000 (0:00:02.751)       0:00:49.009 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Running placement bootstrap container] ***********************\nTuesday 13 January 2026  01:00:22 +0000 (0:00:02.211)       0:00:51.221 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:36 +0000 (0:00:13.736)       0:01:04.957 ******* \n\nTASK [placement : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:36 +0000 (0:00:00.126)       0:01:05.084 ******* \n\nTASK [placement : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:36 +0000 (0:00:00.113)       0:01:05.198 ******* \n\nRUNNING HANDLER [placement : Restart placement-api container] ******************\nTuesday 13 January 2026  01:00:36 +0000 (0:00:00.107)       0:01:05.306 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:00:44 +0000 (0:00:07.841)       0:01:13.147 ******* \n=============================================================================== \nplacement : Running placement bootstrap container ---------------------- 13.74s\nplacement : Restart placement-api container ----------------------------- 7.84s\nservice-ks-register : placement | Creating endpoints -------------------- 6.77s\nservice-ks-register : placement | Granting user roles ------------------- 4.46s\nservice-ks-register : placement | Creating users ------------------------ 4.41s\nservice-ks-register : placement | Creating roles ------------------------ 4.35s\nservice-ks-register : placement | Creating services --------------------- 4.23s\nservice-ks-register : placement | Creating projects --------------------- 3.51s\nplacement : Copying over placement.conf --------------------------------- 3.36s\nplacement : Creating placement databases -------------------------------- 2.75s\nplacement : Creating placement databases user and setting permissions --- 2.21s\nplacement : Copying over migrate-db.rc.j2 configuration ----------------- 1.73s\nservice-cert-copy : placement | Copying over extra CA certificates ------ 1.68s\nplacement : Copying over placement-api wsgi configuration --------------- 1.68s\nplacement : Copying over config.json files for services ----------------- 1.64s\nservice-cert-copy : placement | Copying over backend internal TLS key --- 1.54s\nplacement : Ensuring config directories exist --------------------------- 1.28s\nplacement : Check placement containers ---------------------------------- 1.02s\nservice-cert-copy : placement | Copying over backend internal TLS certificate --- 0.94s\nplacement : include_tasks ----------------------------------------------- 0.69s\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:00:49 +0000 (0:00:00.201)       0:00:00.201 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:00:49 +0000 (0:00:00.295)       0:00:00.496 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Wait for the Keystone service] *******************************************\n\nTASK [Waiting for Keystone public port to be UP] *******************************\nTuesday 13 January 2026  01:00:50 +0000 (0:00:00.643)       0:00:01.139 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:00:51 +0000 (0:00:00.704)       0:00:01.844 ******* \n=============================================================================== \nWaiting for Keystone public port to be UP ------------------------------- 0.70s\nGroup hosts based on enabled services ----------------------------------- 0.64s\nGroup hosts based on Kolla action --------------------------------------- 0.30s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.296)       0:00:00.296 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.318)       0:00:00.615 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m\n\nPLAY [Apply role designate] ****************************************************\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.510)       0:00:01.125 ******* \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : designate | Creating services] *********************\nTuesday 13 January 2026  00:58:24 +0000 (0:00:00.507)       0:00:01.633 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m\n\nTASK [service-ks-register : designate | Creating endpoints] ********************\nTuesday 13 January 2026  00:58:28 +0000 (0:00:04.383)       0:00:06.016 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m\n\nTASK [service-ks-register : designate | Creating projects] *********************\nTuesday 13 January 2026  00:58:36 +0000 (0:00:07.449)       0:00:13.466 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating users] ************************\nTuesday 13 January 2026  00:58:39 +0000 (0:00:03.459)       0:00:16.925 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating roles] ************************\nTuesday 13 January 2026  00:58:44 +0000 (0:00:04.981)       0:00:21.907 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : designate | Granting user roles] *******************\nTuesday 13 January 2026  00:58:48 +0000 (0:00:03.586)       0:00:25.494 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m\n\nTASK [designate : Ensuring config directories exist] ***************************\nTuesday 13 January 2026  00:58:52 +0000 (0:00:04.196)       0:00:29.690 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Check if policies shall be overwritten] **********************\nTuesday 13 January 2026  00:58:55 +0000 (0:00:03.512)       0:00:33.203 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Set designate policy file] ***********************************\nTuesday 13 January 2026  00:58:56 +0000 (0:00:00.127)       0:00:33.330 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 13 January 2026  00:58:56 +0000 (0:00:00.288)       0:00:33.619 ******* \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over extra CA certificates] ******\nTuesday 13 January 2026  00:58:57 +0000 (0:00:00.765)       0:00:34.385 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:59:03 +0000 (0:00:06.760)       0:00:41.146 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'reg\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state SUCCESS\u001b[0m\nistry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS key] ***\nTuesday 13 January 2026  00:59:05 +0000 (0:00:01.991)       0:00:43.138 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over config.json files for services] *****************\nTuesday 13 January 2026  00:59:07 +0000 (0:00:02.090)       0:00:45.228 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over designate.conf] *********************************\nTuesday 13 January 2026  00:59:14 +0000 (0:00:06.970)       0:00:52.198 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over pools.yaml] *************************************\nTuesday 13 January 2026  00:59:37 +0000 (0:00:22.158)       0:01:14.357 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\nTASK [designate : Copying over named.conf] *************************************\nTuesday 13 January 2026  00:59:42 +0000 (0:00:05.526)       0:01:19.884 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\nTASK [designate : Copying over rndc.conf] **************************************\nTuesday 13 January 2026  00:59:45 +0000 (0:00:03.049)       0:01:22.933 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over rndc.key] ***************************************\nTuesday 13 January 2026  00:59:49 +0000 (0:00:03.988)       0:01:26.921 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 13 January 2026  00:59:53 +0000 (0:00:03.645)       0:01:30.567 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over existing policy file] ***************************\nTuesday 13 January 2026  00:59:54 +0000 (0:00:00.783)       0:01:31.350 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Check designate containers] **********************************\nTuesday 13 January 2026  00:59:54 +0000 (0:00:00.716)       0:01:32.067 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 13 January 2026  01:00:00 +0000 (0:00:05.229)       0:01:37.296 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Creating Designate databases] ********************************\nTuesday 13 January 2026  01:00:00 +0000 (0:00:00.585)       0:01:37.881 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m\n\nTASK [designate : Creating Designate databases user and setting permissions] ***\nTuesday 13 January 2026  01:00:03 +0000 (0:00:02.643)       0:01:40.525 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m\n\nTASK [designate : Running Designate bootstrap container] ***********************\nTuesday 13 January 2026  01:00:05 +0000 (0:00:02.537)       0:01:43.063 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [designate : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:21 +0000 (0:00:15.357)       0:01:58.421 ******* \n\nTASK [designate : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:21 +0000 (0:00:00.063)       0:01:58.484 ******* \n\nTASK [designate : Flush handlers] **********************************************\nTuesday 13 January 2026  01:00:21 +0000 (0:00:00.073)       0:01:58.558 ******* \n\nRUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********\nTuesday 13 January 2026  01:00:21 +0000 (0:00:00.069)       0:01:58.627 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-api container] ******************\nTuesday 13 January 2026  01:00:34 +0000 (0:00:13.089)       0:02:11.717 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-central container] **************\nTuesday 13 January 2026  01:00:41 +0000 (0:00:06.992)       0:02:18.709 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-producer container] *************\nTuesday 13 January 2026  01:00:51 +0000 (0:00:09.724)       0:02:28.434 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-mdns container] *****************\nTuesday 13 January 2026  01:00:56 +0000 (0:00:05.027)       0:02:33.461 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-worker container] ***************\nTuesday 13 January 2026  01:01:01 +0000 (0:00:04.871)       0:02:38.333 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [designate : Non-destructive DNS pools update] ****************************\nTuesday 13 January 2026  01:01:07 +0000 (0:00:06.020)       0:02:44.353 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:01:14 +0000 (0:00:07.211)       0:02:51.565 ******* \n=============================================================================== \ndesignate : Copying over designate.conf -------------------------------- 22.16s\ndesignate : Running Designate bootstrap container ---------------------- 15.36s\ndesignate : Restart designate-backend-bind9 container ------------------ 13.09s\ndesignate : Restart designate-central container ------------------------- 9.72s\nservice-ks-register : designate | Creating endpoints -------------------- 7.45s\ndesignate : Non-destructive DNS pools update ---------------------------- 7.21s\ndesignate : Restart designate-api container ----------------------------- 6.99s\ndesignate : Copying over config.json files for services ----------------- 6.97s\nservice-cert-copy : designate | Copying over extra CA certificates ------ 6.76s\ndesignate : Restart designate-worker container -------------------------- 6.02s\ndesignate : Copying over pools.yaml ------------------------------------- 5.53s\ndesignate : Check designate containers ---------------------------------- 5.23s\ndesignate : Restart designate-producer container ------------------------ 5.03s\nservice-ks-register : designate | Creating users ------------------------ 4.98s\ndesignate : Restart designate-mdns container ---------------------------- 4.87s\nservice-ks-register : designate | Creating services --------------------- 4.38s\nservice-ks-register : designate | Granting user roles ------------------- 4.20s\ndesignate : Copying over rndc.conf -------------------------------------- 3.99s\ndesignate : Copying over rndc.key --------------------------------------- 3.65s\nservice-ks-register : designate | Creating roles ------------------------ 3.59s\n\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:01:20 +0000 (0:00:00.464)       0:00:00.464 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:01:22 +0000 (0:00:01.687)       0:00:02.151 ******* \n\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-rgw : include_tasks] ************************************************\nTuesday 13 January 2026  01:01:24 +0000 (0:00:02.389)       0:00:04.540 ******* \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating services] **********************\nTuesday 13 January 2026  01:01:29 +0000 (0:00:05.171)       0:00:09.712 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating endpoints] *********************\nTuesday 13 January 2026  01:01:33 +0000 (0:00:03.719)       0:00:13.433 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating projects] **********************\nTuesday 13 January 2026  01:01:40 +0000 (0:00:06.831)       0:00:20.265 ******* \n\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating users] *************************\nTuesday 13 January 2026  01:01:42 +0000 (0:00:02.703)       0:00:22.969 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating roles] *************************\nTuesday 13 January 2026  01:01:46 +0000 (0:00:03.351)       0:00:26.320 ******* \n\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Granting user roles] ********************\nTuesday 13 January 2026  01:01:52 +0000 (0:00:05.996)       0:00:32.316 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:01:57 +0000 (0:00:05.406)       0:00:37.723 ******* \n=============================================================================== \nservice-ks-register : ceph-rgw | Creating endpoints --------------------- 6.83s\nservice-ks-register : ceph-rgw | Creating roles ------------------------- 6.00s\nservice-ks-register : ceph-rgw | Granting user roles -------------------- 5.41s\nceph-rgw : include_tasks ------------------------------------------------ 5.17s\nservice-ks-register : ceph-rgw | Creating services ---------------------- 3.72s\nservice-ks-register : ceph-rgw | Creating users ------------------------- 3.35s\nservice-ks-register : ceph-rgw | Creating projects ---------------------- 2.70s\nGroup hosts based on enabled services ----------------------------------- 2.39s\nGroup hosts based on Kolla action --------------------------------------- 1.69s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:00:27 +0000 (0:00:00.542)       0:00:00.542 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:00:28 +0000 (0:00:00.329)       0:00:00.871 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m\n\nPLAY [Apply role magnum] *******************************************************\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 13 January 2026  01:00:28 +0000 (0:00:00.381)       0:00:01.253 ******* \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : magnum | Creating services] ************************\nTuesday 13 January 2026  01:00:29 +0000 (0:00:00.604)       0:00:01.858 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m\n\nTASK [service-ks-register : magnum | Creating endpoints] ***********************\nTuesday 13 January 2026  01:00:33 +0000 (0:00:03.891)       0:00:05.749 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating projects] ************************\nTuesday 13 January 2026  01:00:39 +0000 (0:00:06.759)       0:00:12.508 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating users] ***************************\nTuesday 13 January 2026  01:00:42 +0000 (0:00:03.073)       0:00:15.581 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating roles] ***************************\nTuesday 13 January 2026  01:00:46 +0000 (0:00:03.983)       0:00:19.565 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : magnum | Granting user roles] **********************\nTuesday 13 January 2026  01:00:50 +0000 (0:00:03.232)       0:00:22.797 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m\n\nTASK [magnum : Creating Magnum trustee domain] *********************************\nTuesday 13 January 2026  01:00:53 +0000 (0:00:03.616)       0:00:26.414 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user] ***********************************\nTuesday 13 January 2026  01:00:56 +0000 (0:00:03.250)       0:00:29.665 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user role] ******************************\nTuesday 13 January 2026  01:01:00 +0000 (0:00:03.956)       0:00:33.621 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  01:01:04 +0000 (0:00:03.186)       0:00:36.807 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Check if policies shall be overwritten] *************************\nTuesday 13 January 2026  01:01:05 +0000 (0:00:01.865)       0:00:38.673 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Set magnum policy file] *****************************************\nTuesday 13 January 2026  01:01:06 +0000 (0:00:00.125)       0:00:38.798 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check if kubeconfig file is supplied] ***************************\nTuesday 13 January 2026  01:01:06 +0000 (0:00:00.507)       0:00:39.305 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [magnum : Copying over kubeconfig file] ***********************************\nTuesday 13 January 2026  01:01:07 +0000 (0:00:00.884)       0:00:40.191 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Set magnum kubeconfig file's path] ******************************\nTuesday 13 January 2026  01:01:09 +0000 (0:00:02.464)       0:00:42.655 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 13 January 2026  01:01:10 +0000 (0:00:00.289)       0:00:42.945 ******* \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over extra CA certificates] *********\nTuesday 13 January 2026  01:01:11 +0000 (0:00:00.892)       0:00:43.837 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:01:13 +0000 (0:00:02.435)       0:00:46.273 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******\nTuesday 13 January 2026  01:01:14 +0000 (0:00:01.132)       0:00:47.406 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Copying over config.json files for services] ********************\nTuesday 13 January 2026  01:01:16 +0000 (0:00:01.693)       0:00:49.099 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over magnum.conf] ***************************************\nTuesday 13 January 2026  01:01:19 +0000 (0:00:02.940)       0:00:52.040 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over existing policy file] ******************************\nTuesday 13 January 2026  01:01:28 +0000 (0:00:09.650)       0:01:01.691 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check magnum containers] ****************************************\nTuesday 13 January 2026  01:01:30 +0000 (0:00:01.406)       0:01:03.098 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 13 January 2026  01:01:35 +0000 (0:00:04.684)       0:01:07.782 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database] ***************************************\nTuesday 13 January 2026  01:01:36 +0000 (0:00:00.931)       0:01:08.714 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database user and setting permissions] **********\nTuesday 13 January 2026  01:01:38 +0000 (0:00:02.274)       0:01:10.989 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Running Magnum bootstrap container] *****************************\nTuesday 13 January 2026  01:01:40 +0000 (0:00:02.210)       0:01:13.199 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 13 January 2026  01:01:54 +0000 (0:00:13.954)       0:01:27.154 ******* \n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 13 January 2026  01:01:54 +0000 (0:00:00.078)       0:01:27.232 ******* \n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 13 January 2026  01:01:54 +0000 (0:00:00.069)       0:01:27.302 ******* \n\nRUNNING HANDLER [magnum : Restart magnum-api container] ************************\nTuesday 13 January 2026  01:01:54 +0000 (0:00:00.072)       0:01:27.374 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [magnum : Restart magnum-conductor container] ******************\nTuesday 13 January 2026  01:02:09 +0000 (0:00:14.827)       0:01:42.201 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:02:19 +0000 (0:00:09.627)       0:01:51.829 ******* \n=============================================================================== \nmagnum : Restart magnum-api container ---------------------------------- 14.83s\nmagnum : Running Magnum bootstrap container ---------------------------- 13.95s\nmagnum : Copying over magnum.conf --------------------------------------- 9.65s\nmagnum : Restart magnum-conductor container ----------------------------- 9.63s\nservice-ks-register : magnum | Creating endpoints ----------------------- 6.76s\nmagnum : Check magnum containers ---------------------------------------- 4.68s\nservice-ks-register : magnum | Creating users --------------------------- 3.98s\nmagnum : Creating Magnum trustee user ----------------------------------- 3.96s\nservice-ks-register : magnum | Creating services ------------------------ 3.89s\nservice-ks-register : magnum | Granting user roles ---------------------- 3.62s\nmagnum : Creating Magnum trustee domain --------------------------------- 3.25s\nservice-ks-register : magnum | Creating roles --------------------------- 3.23s\nmagnum : Creating Magnum trustee user role ------------------------------ 3.19s\nservice-ks-register : magnum | Creating projects ------------------------ 3.07s\nmagnum : Copying over config.json files for services -------------------- 2.94s\nmagnum : Copying over kubeconfig file ----------------------------------- 2.46s\nservice-cert-copy : magnum | Copying over extra CA certificates --------- 2.44s\nmagnum : Creating Magnum database --------------------------------------- 2.27s\nmagnum : Creating Magnum database user and setting permissions ---------- 2.21s\nmagnum : Ensuring config directories exist ------------------------------ 1.87s\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  00:58:22 +0000 (0:00:00.277)       0:00:00.277 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:01.002)       0:00:01.280 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m\n\nPLAY [Apply role neutron] ******************************************************\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 13 January 2026  00:58:23 +0000 (0:00:00.777)       0:00:02.058 ******* \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [neutron : Get container facts] *******************************************\nTuesday 13 January 2026  00:58:25 +0000 (0:00:01.060)       0:00:03.118 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Get container volume facts] ************************************\nTuesday 13 January 2026  00:58:26 +0000 (0:00:01.212)       0:00:04.331 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check for ML2/OVN presence] ************************************\nTuesday 13 January 2026  00:58:27 +0000 (0:00:01.132)       0:00:05.463 ******* \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [neutron : Check for ML2/OVS presence] ************************************\nTuesday 13 January 2026  00:58:28 +0000 (0:00:00.807)       0:00:06.270 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-ks-register : neutron | Creating services] ***********************\nTuesday 13 January 2026  00:58:28 +0000 (0:00:00.554)       0:00:06.825 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m\n\nTASK [service-ks-register : neutron | Creating endpoints] **********************\nTuesday 13 January 2026  00:58:32 +0000 (0:00:03.627)       0:00:10.453 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating projects] ***********************\nTuesday 13 January 2026  00:58:39 +0000 (0:00:07.329)       0:00:17.782 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating users] **************************\nTuesday 13 January 2026  00:58:43 +0000 (0:00:03.512)       0:00:21.295 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating roles] **************************\nTuesday 13 January 2026  00:58:47 +0000 (0:00:04.422)       0:00:25.717 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : neutron | Granting user roles] *********************\nTuesday 13 January 2026  00:58:51 +0000 (0:00:03.684)       0:00:29.401 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 13 January 2026  00:58:59 +0000 (0:00:07.706)       0:00:37.108 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Load and persist kernel modules] *****************************************\nTuesday 13 January 2026  00:58:59 +0000 (0:00:00.821)       0:00:37.930 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check IPv6 support] ********************************************\nTuesday 13 January 2026  00:59:02 +0000 (0:00:02.417)       0:00:40.347 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nTuesday 13 January 2026  00:59:04 +0000 (0:00:02.052)       0:00:42.399 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Ensuring config directories exist] *****************************\nTuesday 13 January 2026  00:59:07 +0000 (0:00:03.364)       0:00:45.764 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Check if extra ml2 plugins exists] *****************************\nTuesday 13 January 2026  00:59:11 +0000 (0:00:03.692)       0:00:49.456 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m\n\u001b[1;35ma directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 13 January 2026  00:59:12 +0000 (0:00:00.753)       0:00:50.210 ******* \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over extra CA certificates] ********\nTuesday 13 January 2026  00:59:13 +0000 (0:00:00.893)       0:00:51.103 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  00:59:15 +0000 (0:00:02.780)       0:00:53.884 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  00:59:18 +0000 (0:00:03.172)       0:00:57.056 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating TLS backend PEM File] *********************************\nTuesday 13 January 2026  00:59:21 +0000 (0:00:02.473)       0:00:59.529 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  00:59:24 +0000 (0:00:02.654)       0:01:02.184 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Set neutron policy file] ***************************************\nTuesday 13 January 2026  00:59:24 +0000 (0:00:00.181)       0:01:02.365 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over existing policy file] *****************************\nTuesday 13 January 2026  00:59:25 +0000 (0:00:00.733)       0:01:03.099 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_po\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state SUCCESS\u001b[0m\nrt neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over config.json files for services] *******************\nTuesday 13 January 2026  00:59:28 +0000 (0:00:03.295)       0:01:06.395 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron.conf] *************************************\nTuesday 13 January 2026  00:59:32 +0000 (0:00:04.234)       0:01:10.629 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\nTASK [neutron : Copying over neutron_vpnaas.conf] ******************************\nTuesday 13 January 2026  00:59:39 +0000 (0:00:06.942)       0:01:17.571 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over ssh key] ******************************************\nTuesday 13 January 2026  00:59:42 +0000 (0:00:02.536)       0:01:20.108 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [neutron : Copying over ml2_conf.ini] *************************************\nTuesday 13 January 2026  00:59:44 +0000 (0:00:02.896)       0:01:23.005 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\nTASK [neutron : Copying over linuxbridge_agent.ini] ****************************\nTuesday 13 January 2026  00:59:49 +0000 (0:00:04.323)       0:01:27.328 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over openvswitch_agent.ini] ****************************\nTuesday 13 January 2026  00:59:52 +0000 (0:00:02.794)       0:01:30.122 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over sriov_agent.ini] **********************************\nTuesday 13 January 2026  00:59:54 +0000 (0:00:02.490)       0:01:32.613 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over mlnx_agent.ini] ***********************************\nTuesday 13 January 2026  00:59:57 +0000 (0:00:02.603)       0:01:35.216 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over eswitchd.conf] ************************************\nTuesday 13 January 2026  00:59:58 +0000 (0:00:01.776)       0:01:36.993 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over dhcp_agent.ini] ***********************************\nTuesday 13 January 2026  01:00:01 +0000 (0:00:02.683)       0:01:39.677 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over dnsmasq.conf] *************************************\nTuesday 13 January 2026  01:00:03 +0000 (0:00:02.233)       0:01:41.910 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over l3_agent.ini] *************************************\nTuesday 13 January 2026  01:00:05 +0000 (0:00:02.159)       0:01:44.070 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over fwaas_driver.ini] *********************************\nTuesday 13 January 2026  01:00:08 +0000 (0:00:02.912)       0:01:46.983 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', \u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n'/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metadata_agent.ini] *******************************\nTuesday 13 January 2026  01:00:11 +0000 (0:00:02.437)       0:01:49.421 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************\nTuesday 13 January 2026  01:00:13 +0000 (0:00:02.472)       0:01:51.893 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metering_agent.ini] *******************************\nTuesday 13 January 2026  01:00:17 +0000 (0:00:03.433)       0:01:55.326 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ironic_neutron_agent.ini] *************************\nTuesday 13 January 2026  01:00:19 +0000 (0:00:01.857)       0:01:57.183 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over bgp_dragent.ini] **********************************\nTuesday 13 January 2026  01:00:21 +0000 (0:00:02.222)       0:01:59.406 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over ovn_agent.ini] ************************************\nTuesday 13 January 2026  01:00:24 +0000 (0:00:03.220)       0:02:02.626 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over nsx.ini] ******************************************\nTuesday 13 January 2026  01:00:26 +0000 (0:00:02.128)       0:02:04.755 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copy neutron-l3-agent-wrapper script] **************************\nTuesday 13 January 2026  01:00:28 +0000 (0:00:02.031)       0:02:06.787 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over extra ml2 plugins] ********************************\nTuesday 13 January 2026  01:00:30 +0000 (0:00:01.902)       0:02:08.689 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************\nTuesday 13 January 2026  01:00:32 +0000 (0:00:01.826)       0:02:10.515 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over neutron_taas.conf] ********************************\nTuesday 13 January 2026  01:00:35 +0000 (0:00:02.715)       0:02:13.231 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Check neutron containers] **************************************\nTuesday 13 January 2026  01:00:37 +0000 (0:00:02.798)       0:02:16.030 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 13 January 2026  01:00:40 +0000 (0:00:02.568)       0:02:18.598 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating Neutron database] *************************************\nTuesday 13 January 2026  01:00:40 +0000 (0:00:00.458)       0:02:19.056 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Creating Neutron database user and setting permissions] ********\nTuesday 13 January 2026  01:00:42 +0000 (0:00:01.974)       0:02:21.031 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Running Neutron bootstrap container] ***************************\nTuesday 13 January 2026  01:00:45 +0000 (0:00:02.304)       0:02:23.335 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:24 +0000 (0:00:39.430)       0:03:02.766 ******* \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:24 +0000 (0:00:00.185)       0:03:02.952 ******* \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:25 +0000 (0:00:00.237)       0:03:03.190 ******* \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:25 +0000 (0:00:00.130)       0:03:03.321 ******* \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:25 +0000 (0:00:00.055)       0:03:03.376 ******* \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 13 January 2026  01:01:25 +0000 (0:00:00.099)       0:03:03.476 ******* \n\nRUNNING HANDLER [neutron : Restart neutron-server container] *******************\nTuesday 13 January 2026  01:01:25 +0000 (0:00:00.140)       0:03:03.617 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******\nTuesday 13 January 2026  01:01:53 +0000 (0:00:28.430)       0:03:32.047 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:02:45 +0000 (0:00:51.904)       0:04:23.951 ******* \n=============================================================================== \nneutron : Restart neutron-ovn-metadata-agent container ----------------- 51.90s\nneutron : Running Neutron bootstrap container -------------------------- 39.43s\nneutron : Restart neutron-server container ----------------------------- 28.43s\nservice-ks-register : neutron | Granting user roles --------------------- 7.71s\nservice-ks-register : neutron | Creating endpoints ---------------------- 7.33s\nneutron : Copying over neutron.conf ------------------------------------- 6.94s\nservice-ks-register : neutron | Creating users -------------------------- 4.42s\nneutron : Copying over ml2_conf.ini ------------------------------------- 4.32s\nneutron : Copying over config.json files for services ------------------- 4.24s\nneutron : Ensuring config directories exist ----------------------------- 3.69s\nservice-ks-register : neutron | Creating roles -------------------------- 3.68s\nservice-ks-register : neutron | Creating services ----------------------- 3.63s\nservice-ks-register : neutron | Creating projects ----------------------- 3.51s\nneutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 3.43s\nSetting sysctl values --------------------------------------------------- 3.36s\nneutron : Copying over existing policy file ----------------------------- 3.30s\nneutron : Copying over bgp_dragent.ini ---------------------------------- 3.22s\nservice-cert-copy : neutron | Copying over backend internal TLS certificate --- 3.17s\nneutron : Copying over l3_agent.ini ------------------------------------- 2.91s\nneutron : Copying over ssh key ------------------------------------------ 2.90s\n\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:00:55 +0000 (0:00:00.209)       0:00:00.209 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:00:56 +0000 (0:00:00.596)       0:00:00.806 ******* \n\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m\n\nPLAY [Apply role prometheus] ***************************************************\n\nTASK [prometheus : include_tasks] **********************************************\nTuesday 13 January 2026  01:00:56 +0000 (0:00:00.602)       0:00:01.408 ******* \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [prometheus : Ensuring config directories exist] **************************\nTuesday 13 January 2026  01:00:58 +0000 (0:00:01.257)       0:00:02.665 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : include_tasks] **********************************************\nTuesday 13 January 2026  01:01:00 +0000 (0:00:02.649)       0:00:05.315 ******* \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****\nTuesday 13 January 2026  01:01:02 +0000 (0:00:01.435)       0:00:06.751 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:01:08 +0000 (0:00:06.126)       0:00:12.877 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***\nTuesday 13 January 2026  01:01:09 +0000 (0:00:01.526)       0:00:14.403 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over config.json files] *****************************\nTuesday 13 January 2026  01:01:12 +0000 (0:00:02.295)       0:00:16.699 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Find custom prometheus alert rules files] *******************\nTuesday 13 January 2026  01:01:18 +0000 (0:00:06.763)       0:00:23.463 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom prometheus alert rules files] ***********\nTuesday 13 January 2026  01:01:20 +0000 (0:00:01.770)       0:00:25.234 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [prometheus : Find prometheus common config overrides] ********************\nTuesday 13 January 2026  01:01:49 +0000 (0:00:28.699)       0:00:53.933 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Find prometheus host config overrides] **********************\nTuesday 13 January 2026  01:01:50 +0000 (0:00:00.720)       0:00:54.654 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over prometheus config file] ************************\nTuesday 13 January 2026  01:01:52 +0000 (0:00:02.029)       0:00:56.684 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus web config file] ********************\nTuesday 13 January 2026  01:02:11 +0000 (0:00:19.058)       0:01:15.742 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus alertmanager config file] ***********\nTuesday 13 January 2026  01:02:14 +0000 (0:00:03.602)       0:01:19.345 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m\n\nTASK [prometheus : Find custom Alertmanager alert notification templates] ******\nTuesday 13 January 2026  01:02:16 +0000 (0:00:01.424)       0:01:20.769 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom Alertmanager alert notification templates] ***\nTuesday 13 January 2026  01:02:16 +0000 (0:00:00.620)       0:01:21.390 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************\nTuesday 13 January 2026  01:02:17 +0000 (0:00:00.580)       0:01:21.970 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [prometheus : Copying cloud config file for openstack exporter] ***********\nTuesday 13 January 2026  01:02:19 +0000 (0:00:01.764)       0:01:23.734 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying config file for blackbox exporter] ******************\nTuesday 13 January 2026  01:02:21 +0000 (0:00:01.857)       0:01:25.592 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Find extra prometheus server config files] ******************\nTuesday 13 January 2026  01:02:22 +0000 (0:00:01.575)       0:01:27.168 ******* \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m\n\u001b[1;35mnot a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Create subdirectories for extra config files] ***************\nTuesday 13 January 2026  01:02:23 +0000 (0:00:01.011)       0:01:28.179 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Template extra prometheus server config files] **************\nTuesday 13 January 2026  01:02:24 +0000 (0:00:00.705)       0:01:28.885 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Check prometheus containers] ********************************\nTuesday 13 January 2026  01:02:25 +0000 (0:00:00.672)       0:01:29.558 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Creating prometheus database user and setting permissions] ***\nTuesday 13 January 2026  01:02:28 +0000 (0:00:03.618)       0:01:33.177 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:29 +0000 (0:00:01.319)       0:01:34.497 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.081)       0:01:34.578 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.064)       0:01:34.643 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.059)       0:01:34.702 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.225)       0:01:34.928 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.063)       0:01:34.991 ******* \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.063)       0:01:35.055 ******* \n\nRUNNING HANDLER [prometheus : Restart prometheus-server container] *************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.085)       0:01:35.141 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******\nTuesday 13 January 2026  01:02:47 +0000 (0:00:16.704)       0:01:51.846 ******* \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****\nTuesday 13 January 2026  01:03:01 +0000 (0:00:14.270)       0:02:06.116 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***\nTuesday 13 January 2026  01:03:11 +0000 (0:00:10.338)       0:02:16.455 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********\nTuesday 13 January 2026  01:03:24 +0000 (0:00:12.314)       0:02:28.769 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******\nTuesday 13 January 2026  01:03:38 +0000 (0:00:14.651)       0:02:43.420 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***\nTuesday 13 January 2026  01:03:47 +0000 (0:00:08.192)       0:02:51.613 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***\nTuesday 13 January 2026  01:03:51 +0000 (0:00:04.742)       0:02:56.356 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***\nTuesday 13 January 2026  01:03:56 +0000 (0:00:04.951)       0:03:01.308 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:04:06 +0000 (0:00:09.826)       0:03:11.134 ******* \n=============================================================================== \nprometheus : Copying over custom prometheus alert rules files ---------- 28.70s\nprometheus : Copying over prometheus config file ----------------------- 19.06s\nprometheus : Restart prometheus-server container ----------------------- 16.70s\nprometheus : Restart prometheus-cadvisor container --------------------- 14.65s\nprometheus : Restart prometheus-node-exporter container ---------------- 14.27s\nprometheus : Restart prometheus-memcached-exporter container ----------- 12.31s\nprometheus : Restart prometheus-mysqld-exporter container -------------- 10.34s\nprometheus : Restart prometheus-libvirt-exporter container -------------- 9.83s\nprometheus : Restart prometheus-alertmanager container ------------------ 8.19s\nprometheus : Copying over config.json files ----------------------------- 6.76s\nservice-cert-copy : prometheus | Copying over extra CA certificates ----- 6.13s\nprometheus : Restart prometheus-blackbox-exporter container ------------- 4.95s\nprometheus : Restart prometheus-elasticsearch-exporter container -------- 4.74s\nprometheus : Check prometheus containers -------------------------------- 3.62s\nprometheus : Copying over prometheus web config file -------------------- 3.60s\nprometheus : Ensuring config directories exist -------------------------- 2.65s\nservice-cert-copy : prometheus | Copying over backend internal TLS key --- 2.30s\nprometheus : Find prometheus host config overrides ---------------------- 2.03s\nprometheus : Copying cloud config file for openstack exporter ----------- 1.86s\nprometheus : Find custom prometheus alert rules files ------------------- 1.77s\n\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:02:02 +0000 (0:00:00.245)       0:00:00.245 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:02:03 +0000 (0:00:00.251)       0:00:00.496 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m\n\nPLAY [Apply role glance] *******************************************************\n\nTASK [glance : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:03 +0000 (0:00:00.343)       0:00:00.840 ******* \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : glance | Creating services] ************************\nTuesday 13 January 2026  01:02:03 +0000 (0:00:00.446)       0:00:01.286 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m\n\nTASK [service-ks-register : glance | Creating endpoints] ***********************\nTuesday 13 January 2026  01:02:08 +0000 (0:00:04.079)       0:00:05.365 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m\n\nTASK [service-ks-register : glance | Creating projects] ************************\nTuesday 13 January 2026  01:02:14 +0000 (0:00:06.159)       0:00:11.525 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating users] ***************************\nTuesday 13 January 2026  01:02:17 +0000 (0:00:02.823)       0:00:14.348 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating roles] ***************************\nTuesday 13 January 2026  01:02:20 +0000 (0:00:03.436)       0:00:17.784 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : glance | Granting user roles] **********************\nTuesday 13 January 2026  01:02:23 +0000 (0:00:03.059)       0:00:20.844 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m\n\nTASK [glance : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  01:02:26 +0000 (0:00:03.334)       0:00:24.178 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:03.420)       0:00:27.599 ******* \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [glance : Ensuring glance service ceph config subdir exists] **************\nTuesday 13 January 2026  01:02:30 +0000 (0:00:00.663)       0:00:28.262 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copy over multiple ceph configs for Glance] *********************\nTuesday 13 January 2026  01:02:34 +0000 (0:00:03.638)       0:00:31.901 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Copy over ceph Glance keyrings] *********************************\nTuesday 13 January 2026  01:02:36 +0000 (0:00:01.656)       0:00:33.557 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Ensuring config directory has correct owner and permission] *****\nTuesday 13 January 2026  01:02:37 +0000 (0:00:01.292)       0:00:34.850 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check if policies shall be overwritten] *************************\nTuesday 13 January 2026  01:02:38 +0000 (0:00:00.648)       0:00:35.499 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Set glance policy file] *****************************************\nTuesday 13 January 2026  01:02:38 +0000 (0:00:00.375)       0:00:35.875 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:38 +0000 (0:00:00.291)       0:00:36.166 ******* \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over extra CA certificates] *********\nTuesday 13 January 2026  01:02:39 +0000 (0:00:00.531)       0:00:36.698 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:02:43 +0000 (0:00:04.540)       0:00:41.239 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS key] ******\nTuesday 13 January 2026  01:02:48 +0000 (0:00:04.068)       0:00:45.307 ******* \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating TLS backend PEM File] **********************************\nTuesday 13 January 2026  01:02:54 +0000 (0:00:06.092)       0:00:51.399 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over config.json files for services] ********************\nTuesday 13 January 2026  01:02:58 +0000 (0:00:04.250)       0:00:55.650 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : Copying over glance-api.conf] ***********************************\nTuesday 13 January 2026  01:03:02 +0000 (0:00:03.935)       0:00:59.585 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over glance-cache.conf for glance_api] ******************\nTuesday 13 January 2026  01:03:08 +0000 (0:00:05.897)       0:01:05.482 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-swift.conf for glance_api] ******************\nTuesday 13 January 2026  01:03:11 +0000 (0:00:03.500)       0:01:08.983 ******* \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over glance-image-import.conf] **************************\nTuesday 13 January 2026  01:03:16 +0000 (0:00:04.952)       0:01:13.935 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over property-protections-rules.conf] *******************\nTuesday 13 January 2026  01:03:19 +0000 (0:00:03.028)       0:01:16.963 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over existing policy file] ******************************\nTuesday 13 January 2026  01:03:22 +0000 (0:00:03.011)       0:01:19.974 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-haproxy-tls.cfg] ****************************\nTuesday 13 January 2026  01:03:22 +0000 (0:00:00.261)       0:01:20.236 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostnqn' file for glance_api] ***********************\nTuesday 13 January 2026  01:03:29 +0000 (0:00:06.723)       0:01:26.959 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check glance containers] ****************************************\nTuesday 13 January 2026  01:03:35 +0000 (0:00:05.818)       0:01:32.778 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 13 January 2026  01:03:38 +0000 (0:00:03.476)       0:01:36.255 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating Glance database] ***************************************\nTuesday 13 January 2026  01:03:39 +0000 (0:00:00.254)       0:01:36.509 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating Glance database user and setting permissions] **********\nTuesday 13 January 2026  01:03:41 +0000 (0:00:02.055)       0:01:38.565 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Enable log_bin_trust_function_creators function] ****************\nTuesday 13 January 2026  01:03:43 +0000 (0:00:02.452)       0:01:41.017 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Running Glance bootstrap container] *****************************\nTuesday 13 January 2026  01:03:45 +0000 (0:00:02.183)       0:01:43.201 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Disable log_bin_trust_function_creators function] ***************\nTuesday 13 January 2026  01:04:13 +0000 (0:00:27.604)       0:02:10.805 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:15 +0000 (0:00:02.075)       0:02:12.881 ******* \n\nTASK [glance : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:15 +0000 (0:00:00.337)       0:02:13.219 ******* \n\nTASK [glance : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:15 +0000 (0:00:00.070)       0:02:13.289 ******* \n\nRUNNING HANDLER [glance : Restart glance-api container] ************************\nTuesday 13 January 2026  01:04:16 +0000 (0:00:00.082)       0:02:13.372 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:04:45 +0000 (0:00:29.430)       0:02:42.803 ******* \n=============================================================================== \nglance : Restart glance-api container ---------------------------------- 29.43s\nglance : Running Glance bootstrap container ---------------------------- 27.60s\nglance : Copying over glance-haproxy-tls.cfg ---------------------------- 6.72s\nservice-ks-register : glance | Creating endpoints ----------------------- 6.16s\nservice-cert-copy : glance | Copying over backend internal TLS key ------ 6.09s\nglance : Copying over glance-api.conf ----------------------------------- 5.90s\nglance : Generating 'hostnqn' file for glance_api ----------------------- 5.82s\nglance : Copying over glance-swift.conf for glance_api ------------------ 4.95s\nservice-cert-copy : glance | Copying over extra CA certificates --------- 4.54s\nglance : Creating TLS backend PEM File ---------------------------------- 4.25s\nservice-ks-register : glance | Creating services ------------------------ 4.08s\nservice-cert-copy : glance | Copying over backend internal TLS certificate --- 4.07s\nglance : Copying over config.json files for services -------------------- 3.94s\nglance : Ensuring glance service ceph config subdir exists -------------- 3.64s\nglance : Copying over glance-cache.conf for glance_api ------------------ 3.50s\nglance : Check glance containers ---------------------------------------- 3.48s\nservice-ks-register : glance | Creating users --------------------------- 3.44s\nglance : Ensuring config directories exist ------------------------------ 3.42s\nservice-ks-register : glance | Granting user roles ---------------------- 3.33s\nservice-ks-register : glance | Creating roles --------------------------- 3.06s\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:02:23 +0000 (0:00:00.226)       0:00:00.227 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:02:24 +0000 (0:00:00.280)       0:00:00.507 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m\n\nPLAY [Apply role cinder] *******************************************************\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:24 +0000 (0:00:00.346)       0:00:00.853 ******* \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : cinder | Creating services] ************************\nTuesday 13 January 2026  01:02:24 +0000 (0:00:00.480)       0:00:01.334 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m\n\nTASK [service-ks-register : cinder | Creating endpoints] ***********************\nTuesday 13 January 2026  01:02:28 +0000 (0:00:03.247)       0:00:04.581 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating projects] ************************\nTuesday 13 January 2026  01:02:33 +0000 (0:00:05.869)       0:00:10.451 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating users] ***************************\nTuesday 13 January 2026  01:02:37 +0000 (0:00:03.811)       0:00:14.263 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating roles] ***************************\nTuesday 13 January 2026  01:02:41 +0000 (0:00:03.702)       0:00:17.965 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : cinder | Granting user roles] **********************\nTuesday 13 January 2026  01:02:45 +0000 (0:00:03.850)       0:00:21.816 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m\n\nTASK [cinder : Ensuring config directories exist] ******************************\nTuesday 13 January 2026  01:02:52 +0000 (0:00:07.064)       0:00:28.880 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:55 +0000 (0:00:02.696)       0:00:31.577 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 13 January 2026  01:02:55 +0000 (0:00:00.515)       0:00:32.092 ******* \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [cinder : Ensuring cinder service ceph config subdirs exists] *************\nTuesday 13 January 2026  01:02:56 +0000 (0:00:01.128)       0:00:33.221 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Copying over multiple ceph.conf for cinder services] ************\nTuesday 13 January 2026  01:02:58 +0000 (0:00:01.896)       0:00:35.117 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************\nTuesday 13 January 2026  01:03:01 +0000 (0:00:03.239)       0:00:38.357 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************\nTuesday 13 January 2026  01:03:04 +0000 (0:00:02.436)       0:00:40.794 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\nTASK [cinder : Ensuring config directory has correct owner and permission] *****\nTuesday 13 January 2026  01:03:07 +0000 (0:00:03.127)       0:00:43.921 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Check if policies shall be overwritten] *************************\nTuesday 13 January 2026  01:03:08 +0000 (0:00:00.976)       0:00:44.898 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Set cinder policy file] *****************************************\nTuesday 13 January 2026  01:03:08 +0000 (0:00:00.097)       0:00:44.996 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 13 January 2026  01:03:08 +0000 (0:00:00.336)       0:00:45.332 ******* \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over extra CA certificates] *********\nTuesday 13 January 2026  01:03:09 +0000 (0:00:00.646)       0:00:45.979 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:03:14 +0000 (0:00:04.885)       0:00:50.864 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******\nTuesday 13 January 2026  01:03:15 +0000 (0:00:01.128)       0:00:51.993 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over config.json files for services] ********************\nTuesday 13 January 2026  01:03:16 +0000 (0:00:01.347)       0:00:53.341 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Copying over cinder-wsgi.conf] **********************************\nTuesday 13 January 2026  01:03:20 +0000 (0:00:04.119)       0:00:57.460 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\nTASK [cinder : Copying over cinder.conf] ***************************************\nTuesday 13 January 2026  01:03:22 +0000 (0:00:01.838)       0:00:59.299 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************\nTuesday 13 January 2026  01:03:38 +0000 (0:00:15.185)       0:01:14.485 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over existing policy file] ******************************\nTuesday 13 January 2026  01:03:39 +0000 (0:00:01.468)       0:01:15.953 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over nfs_shares files for cinder_volume] ****************\nTuesday 13 January 2026  01:03:40 +0000 (0:00:00.582)       0:01:16.536 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Check cinder containers] ****************************************\nTuesday 13 January 2026  01:03:40 +0000 (0:00:00.289)       0:01:16.825 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 13 January 2026  01:03:43 +0000 (0:00:03.248)       0:01:20.073 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [cinder : Creating Cinder database] ***************************************\nTuesday 13 January 2026  01:03:44 +0000 (0:00:00.697)       0:01:20.771 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Creating Cinder database user and setting permissions] **********\nTuesday 13 January 2026  01:03:46 +0000 (0:00:02.178)       0:01:22.950 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Running Cinder bootstrap container] *****************************\nTuesday 13 January 2026  01:03:48 +0000 (0:00:02.224)       0:01:25.175 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:08 +0000 (0:00:19.633)       0:01:44.808 ******* \n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:08 +0000 (0:00:00.067)       0:01:44.876 ******* \n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 13 January 2026  01:04:08 +0000 (0:00:00.065)       0:01:44.941 ******* \n\nRUNNING HANDLER [cinder : Restart cinder-api container] ************************\nTuesday 13 January 2026  01:04:08 +0000 (0:00:00.064)       0:01:45.005 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************\nTuesday 13 January 2026  01:04:32 +0000 (0:00:24.405)       0:02:09.411 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-volume container] *********************\nTuesday 13 January 2026  01:04:43 +0000 (0:00:10.515)       0:02:19.927 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-backup container] *********************\nTuesday 13 January 2026  01:05:01 +0000 (0:00:17.661)       0:02:37.588 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***\nTuesday 13 January 2026  01:05:06 +0000 (0:00:05.728)       0:02:43.317 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:05:07 +0000 (0:00:00.259)       0:02:43.577 ******* \n=============================================================================== \ncinder : Restart cinder-api container ---------------------------------- 24.41s\ncinder : Running Cinder bootstrap container ---------------------------- 19.63s\ncinder : Restart cinder-volume container ------------------------------- 17.66s\ncinder : Copying over cinder.conf -------------------------------------- 15.19s\ncinder : Restart cinder-scheduler container ---------------------------- 10.52s\nservice-ks-register : cinder | Granting user roles ---------------------- 7.06s\nservice-ks-register : cinder | Creating endpoints ----------------------- 5.87s\ncinder : Restart cinder-backup container -------------------------------- 5.73s\nservice-cert-copy : cinder | Copying over extra CA certificates --------- 4.89s\ncinder : Copying over config.json files for services -------------------- 4.12s\nservice-ks-register : cinder | Creating roles --------------------------- 3.85s\nservice-ks-register : cinder | Creating projects ------------------------ 3.81s\nservice-ks-register : cinder | Creating users --------------------------- 3.70s\ncinder : Check cinder containers ---------------------------------------- 3.25s\nservice-ks-register : cinder | Creating services ------------------------ 3.25s\ncinder : Copying over multiple ceph.conf for cinder services ------------ 3.24s\ncinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.13s\ncinder : Ensuring config directories exist ------------------------------ 2.70s\ncinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.44s\ncinder : Creating Cinder database user and setting permissions ---------- 2.22s\n\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:04:11 +0000 (0:00:00.184)       0:00:00.185 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:04:11 +0000 (0:00:00.347)       0:00:00.532 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Wait for the Nova service] ***********************************************\n\nTASK [Waiting for Nova public port to be UP] ***********************************\nTuesday 13 January 2026  01:04:12 +0000 (0:00:01.181)       0:00:01.714 ******* \n\nSTILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:06:49 +0000 (0:02:36.889)       0:02:38.604 ******* \n=============================================================================== \nWaiting for Nova public port to be UP --------------------------------- 156.89s\nGroup hosts based on enabled services ----------------------------------- 1.18s\nGroup hosts based on Kolla action --------------------------------------- 0.35s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:04:50 +0000 (0:00:00.275)       0:00:00.275 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:04:50 +0000 (0:00:00.318)       0:00:00.594 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m\n\nPLAY [Apply role grafana] ******************************************************\n\nTASK [grafana : include_tasks] *************************************************\nTuesday 13 January 2026  01:04:51 +0000 (0:00:00.443)       0:00:01.038 ******* \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [grafana : Ensuring config directories exist] *****************************\nTuesday 13 January 2026  01:04:51 +0000 (0:00:00.487)       0:00:01.525 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Check if extra configuration file exists] **********************\nTuesday 13 January 2026  01:04:52 +0000 (0:00:00.704)       0:00:02.230 ******* \n\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m\n\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : include_tasks] *************************************************\nTuesday 13 January 2026  01:04:53 +0000 (0:00:00.824)       0:00:03.055 ******* \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over extra CA certificates] ********\nTuesday 13 January 2026  01:04:54 +0000 (0:00:00.678)       0:00:03.733 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:04:55 +0000 (0:00:01.341)       0:00:05.075 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  01:04:55 +0000 (0:00:00.374)       0:00:05.450 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over config.json files] ********************************\nTuesday 13 January 2026  01:04:56 +0000 (0:00:00.823)       0:00:06.273 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over grafana.ini] **************************************\nTuesday 13 January 2026  01:04:57 +0000 (0:00:01.107)       0:00:07.380 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over extra configuration file] *************************\nTuesday 13 January 2026  01:04:58 +0000 (0:00:01.135)       0:00:08.515 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Configuring Prometheus as data source for Grafana] *************\nTuesday 13 January 2026  01:04:59 +0000 (0:00:00.515)       0:00:09.031 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\nTASK [grafana : Configuring dashboards provisioning] ***************************\nTuesday 13 January 2026  01:05:00 +0000 (0:00:01.200)       0:00:10.231 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\nTASK [grafana : Find custom grafana dashboards] ********************************\nTuesday 13 January 2026  01:05:01 +0000 (0:00:01.187)       0:00:11.419 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : Find templated grafana dashboards] *****************************\nTuesday 13 January 2026  01:05:02 +0000 (0:00:00.762)       0:00:12.181 ******* \n\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m\n\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Prune templated Grafana dashboards] ****************************\nTuesday 13 January 2026  01:05:03 +0000 (0:00:00.729)       0:00:12.910 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over custom dashboards] ********************************\nTuesday 13 January 2026  01:05:03 +0000 (0:00:00.516)       0:00:13.427 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\nTASK [grafana : Check grafana containers] **************************************\nTuesday 13 January 2026  01:05:39 +0000 (0:00:35.874)       0:00:49.302 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Creating grafana database] *************************************\nTuesday 13 January 2026  01:05:40 +0000 (0:00:01.050)       0:00:50.352 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Creating grafana database user and setting permissions] ********\nTuesday 13 January 2026  01:05:42 +0000 (0:00:02.302)       0:00:52.654 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 13 January 2026  01:05:45 +0000 (0:00:02.216)       0:00:54.871 ******* \n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 13 January 2026  01:05:45 +0000 (0:00:00.063)       0:00:54.934 ******* \n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 13 January 2026  01:05:45 +0000 (0:00:00.060)       0:00:54.994 ******* \n\nRUNNING HANDLER [grafana : Restart first grafana container] ********************\nTuesday 13 January 2026  01:05:45 +0000 (0:00:00.225)       0:00:55.219 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********\nTuesday 13 January 2026  01:05:47 +0000 (0:00:01.858)       0:00:57.078 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Restart remaining grafana containers] ***************\nTuesday 13 January 2026  01:06:25 +0000 (0:00:38.331)       0:01:35.409 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [grafana : Wait for grafana application ready] ****************************\nTuesday 13 January 2026  01:06:52 +0000 (0:00:26.777)       0:02:02.187 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Remove old grafana docker volume] ******************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:01.950)       0:02:04.138 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Enable grafana datasources] ************************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:00.361)       0:02:04.499 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m\n\nTASK [grafana : Disable Getting Started panel] *********************************\nTuesday 13 January 2026  01:06:57 +0000 (0:00:02.545)       0:02:07.045 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:06:57 +0000 (0:00:00.244)       0:02:07.289 ******* \n=============================================================================== \ngrafana : Waiting for grafana to start on first node ------------------- 38.33s\ngrafana : Copying over custom dashboards ------------------------------- 35.87s\ngrafana : Restart remaining grafana containers ------------------------- 26.78s\ngrafana : Enable grafana datasources ------------------------------------ 2.55s\ngrafana : Creating grafana database ------------------------------------- 2.30s\ngrafana : Creating grafana database user and setting permissions -------- 2.22s\ngrafana : Wait for grafana application ready ---------------------------- 1.95s\ngrafana : Restart first grafana container ------------------------------- 1.86s\nservice-cert-copy : grafana | Copying over extra CA certificates -------- 1.34s\ngrafana : Configuring Prometheus as data source for Grafana ------------- 1.20s\ngrafana : Configuring dashboards provisioning --------------------------- 1.19s\ngrafana : Copying over grafana.ini -------------------------------------- 1.14s\ngrafana : Copying over config.json files -------------------------------- 1.11s\ngrafana : Check grafana containers -------------------------------------- 1.05s\ngrafana : Check if extra configuration file exists ---------------------- 0.82s\nservice-cert-copy : grafana | Copying over backend internal TLS key ----- 0.82s\ngrafana : Find custom grafana dashboards -------------------------------- 0.76s\ngrafana : Find templated grafana dashboards ----------------------------- 0.73s\ngrafana : Ensuring config directories exist ----------------------------- 0.70s\ngrafana : include_tasks ------------------------------------------------- 0.68s\n\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on OpenStack release] **********************************\nTuesday 13 January 2026  01:02:54 +0000 (0:00:00.481)       0:00:00.481 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:02:56 +0000 (0:00:01.503)       0:00:01.985 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:02:56 +0000 (0:00:00.858)       0:00:02.843 ******* \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Bootstrap nova API databases] ********************************************\n\nTASK [Bootstrap deploy] ********************************************************\nTuesday 13 January 2026  01:02:57 +0000 (0:00:00.933)       0:00:03.777 ******* \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Creating Nova databases] ******************************************\nTuesday 13 January 2026  01:02:58 +0000 (0:00:00.737)       0:00:04.514 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m\n\nTASK [nova : Creating Nova databases user and setting permissions] *************\nTuesday 13 January 2026  01:03:02 +0000 (0:00:03.999)       0:00:08.514 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nTuesday 13 January 2026  01:03:07 +0000 (0:00:04.448)       0:00:12.963 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over config.json files for nova-api-bootstrap] ************\nTuesday 13 January 2026  01:03:07 +0000 (0:00:00.589)       0:00:13.552 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************\nTuesday 13 January 2026  01:03:08 +0000 (0:00:01.133)       0:00:14.686 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 13 January 2026  01:03:11 +0000 (0:00:02.835)       0:00:17.521 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nTuesday 13 January 2026  01:03:12 +0000 (0:00:00.344)       0:00:17.866 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Create cell0 mappings] ********************************************\nTuesday 13 January 2026  01:03:43 +0000 (0:00:31.023)       0:00:48.889 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 13 January 2026  01:03:58 +0000 (0:00:15.495)       0:01:04.384 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 13 January 2026  01:04:10 +0000 (0:00:11.758)       0:01:16.143 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Update cell0 mappings] ********************************************\nTuesday 13 January 2026  01:04:11 +0000 (0:00:01.248)       0:01:17.392 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 13 January 2026  01:04:12 +0000 (0:00:00.530)       0:01:17.922 ******* \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nTuesday 13 January 2026  01:04:12 +0000 (0:00:00.488)       0:01:18.411 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nTuesday 13 January 2026  01:04:30 +0000 (0:00:18.251)       0:01:36.663 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Bootstrap nova cell databases] *******************************************\n\nTASK [Bootstrap deploy] ********************************************************\nTuesday 13 January 2026  01:04:31 +0000 (0:00:00.249)       0:01:36.913 ******* \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database] *********************************\nTuesday 13 January 2026  01:04:31 +0000 (0:00:00.442)       0:01:37.355 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database user and setting permissions] ****\nTuesday 13 January 2026  01:04:33 +0000 (0:00:02.014)       0:01:39.370 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nTuesday 13 January 2026  01:04:35 +0000 (0:00:02.284)       0:01:41.654 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nTuesday 13 January 2026  01:04:36 +0000 (0:00:00.361)       0:01:42.016 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nTuesday 13 January 2026  01:04:43 +0000 (0:00:06.992)       0:01:49.009 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nTuesday 13 January 2026  01:04:43 +0000 (0:00:00.321)       0:01:49.331 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nTuesday 13 January 2026  01:04:44 +0000 (0:00:00.972)       0:01:50.303 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******\nTuesday 13 January 2026  01:04:45 +0000 (0:00:00.817)       0:01:51.121 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************\nTuesday 13 January 2026  01:04:46 +0000 (0:00:00.944)       0:01:52.065 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Running Nova cell bootstrap container] ***********************\nTuesday 13 January 2026  01:04:48 +0000 (0:00:02.147)       0:01:54.213 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 13 January 2026  01:05:09 +0000 (0:00:20.709)       0:02:14.922 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 13 January 2026  01:05:21 +0000 (0:00:12.434)       0:02:27.356 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Create cell] *************************************************\nTuesday 13 January 2026  01:05:22 +0000 (0:00:00.840)       0:02:28.197 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Update cell] *************************************************\nTuesday 13 January 2026  01:05:36 +0000 (0:00:13.770)       0:02:41.968 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nTuesday 13 January 2026  01:05:37 +0000 (0:00:01.125)       0:02:43.093 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova] *********************************************************\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 13 January 2026  01:05:37 +0000 (0:00:00.583)       0:02:43.677 ******* \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : nova | Creating services] **************************\nTuesday 13 January 2026  01:05:38 +0000 (0:00:00.550)       0:02:44.227 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m\n\nTASK [service-ks-register : nova | Creating endpoints] *************************\nTuesday 13 January 2026  01:05:41 +0000 (0:00:03.403)       0:02:47.630 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m\n\nTASK [service-ks-register : nova | Creating projects] **************************\nTuesday 13 January 2026  01:05:48 +0000 (0:00:06.635)       0:02:54.266 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating users] *****************************\nTuesday 13 January 2026  01:05:51 +0000 (0:00:02.940)       0:02:57.206 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating roles] *****************************\nTuesday 13 January 2026  01:05:55 +0000 (0:00:03.892)       0:03:01.099 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : nova | Granting user roles] ************************\nTuesday 13 January 2026  01:05:58 +0000 (0:00:03.096)       0:03:04.195 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nTuesday 13 January 2026  01:06:05 +0000 (0:00:07.283)       0:03:11.479 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Check if policies shall be overwritten] ***************************\nTuesday 13 January 2026  01:06:07 +0000 (0:00:01.439)       0:03:12.918 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : Set nova policy file] *********************************************\nTuesday 13 January 2026  01:06:07 +0000 (0:00:00.132)       0:03:13.051 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check for vendordata file] ****************************************\nTuesday 13 January 2026  01:06:07 +0000 (0:00:00.279)       0:03:13.331 ******* \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [nova : Set vendordata file path] *****************************************\nTuesday 13 January 2026  01:06:08 +0000 (0:00:00.860)       0:03:14.191 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 13 January 2026  01:06:08 +0000 (0:00:00.275)       0:03:14.467 ******* \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nTuesday 13 January 2026  01:06:09 +0000 (0:00:00.528)       0:03:14.995 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:06:11 +0000 (0:00:02.601)       0:03:17.596 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nTuesday 13 January 2026  01:06:12 +0000 (0:00:00.601)       0:03:18.198 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over config.json files for services] **********************\nTuesday 13 January 2026  01:06:13 +0000 (0:00:00.806)       0:03:19.005 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over nova.conf] *******************************************\nTuesday 13 January 2026  01:06:16 +0000 (0:00:02.911)       0:03:21.916 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over existing policy file] ********************************\nTuesday 13 January 2026  01:06:21 +0000 (0:00:05.339)       0:03:27.255 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-api-wsgi.conf] **********************************\nTuesday 13 January 2026  01:06:21 +0000 (0:00:00.572)       0:03:27.828 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over vendordata file] *************************************\nTuesday 13 January 2026  01:06:23 +0000 (0:00:01.422)       0:03:29.250 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check nova containers] ********************************************\nTuesday 13 January 2026  01:06:23 +0000 (0:00:00.329)       0:03:29.580 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 13 January 2026  01:06:25 +0000 (0:00:02.118)       0:03:31.699 ******* \n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 13 January 2026  01:06:25 +0000 (0:00:00.139)       0:03:31.839 ******* \n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 13 January 2026  01:06:26 +0000 (0:00:00.135)       0:03:31.974 ******* \n\nRUNNING HANDLER [nova : Restart nova-scheduler container] **********************\nTuesday 13 January 2026  01:06:26 +0000 (0:00:00.152)       0:03:32.126 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-api container] ****************************\nTuesday 13 January 2026  01:06:41 +0000 (0:00:15.330)       0:03:47.456 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role nova-cell] ****************************************************\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 13 January 2026  01:06:52 +0000 (0:00:11.102)       0:03:58.559 ******* \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:01.376)       0:03:59.935 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Load and persist br_netfilter module] ************************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:00.523)       0:04:00.459 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 13 January 2026  01:06:55 +0000 (0:00:00.901)       0:04:01.360 ******* \n\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 13 January 2026  01:06:56 +0000 (0:00:00.602)       0:04:01.963 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 13 January 2026  01:06:57 +0000 (0:00:01.192)       0:04:03.155 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************\nTuesday 13 January 2026  01:06:57 +0000 (0:00:00.443)       0:04:03.599 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\nTASK [nova-cell : Install udev kolla kvm rules] ********************************\nTuesday 13 January 2026  01:06:58 +0000 (0:00:01.225)       0:04:04.824 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Mask qemu-kvm service] ***************************************\nTuesday 13 January 2026  01:07:00 +0000 (0:00:01.134)       0:04:05.958 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nTuesday 13 January 2026  01:07:01 +0000 (0:00:01.787)       0:04:07.746 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 13 January 2026  01:07:03 +0000 (0:00:02.011)       0:04:09.758 ******* \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nTuesday 13 January 2026  01:07:05 +0000 (0:00:01.374)       0:04:11.132 ******* \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:07:08 +0000 (0:00:03.150)       0:04:14.283 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nTuesday 13 January 2026  01:07:10 +0000 (0:00:01.570)       0:04:15.854 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 13 January 2026  01:07:12 +0000 (0:00:02.261)       0:04:18.115 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check nova keyring file] *************************************\nTuesday 13 January 2026  01:07:13 +0000 (0:00:01.053)       0:04:19.169 ******* \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Check cinder keyring file] ***********************************\nTuesday 13 January 2026  01:07:14 +0000 (0:00:00.946)       0:04:20.115 ******* \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Extract nova key from file] **********************************\nTuesday 13 January 2026  01:07:15 +0000 (0:00:00.874)       0:04:20.990 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Extract cinder key from file] ********************************\nTuesday 13 January 2026  01:07:15 +0000 (0:00:00.486)       0:04:21.476 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph nova keyring file] ****************************\nTuesday 13 January 2026  01:07:16 +0000 (0:00:00.759)       0:04:22.236 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph cinder keyring file] **************************\nTuesday 13 January 2026  01:07:17 +0000 (0:00:01.050)       0:04:23.287 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf] *****************************************\nTuesday 13 January 2026  01:07:18 +0000 (0:00:01.071)       0:04:24.358 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m\n\nTASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************\nTuesday 13 January 2026  01:07:21 +0000 (0:00:03.291)       0:04:27.650 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************\nTuesday 13 January 2026  01:07:22 +0000 (0:00:00.545)       0:04:28.195 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring libvirt secrets directory exists] *******************\nTuesday 13 January 2026  01:07:22 +0000 (0:00:00.310)       0:04:28.506 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Pushing nova secret xml for libvirt] *************************\nTuesday 13 January 2026  01:07:23 +0000 (0:00:01.119)       0:04:29.625 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\nTASK [nova-cell : Pushing secrets key for libvirt] *****************************\nTuesday 13 January 2026  01:07:26 +0000 (0:00:03.072)       0:04:32.697 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Check if policies shall be overwritten] **********************\nTuesday 13 January 2026  01:07:29 +0000 (0:00:02.996)       0:04:35.694 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Set nova policy file] ****************************************\nTuesday 13 January 2026  01:07:29 +0000 (0:00:00.126)       0:04:35.821 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check for vendordata file] ***********************************\nTuesday 13 January 2026  01:07:30 +0000 (0:00:00.594)       0:04:36.416 ******* \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\nTASK [nova-cell : Set vendordata file path] ************************************\nTuesday 13 January 2026  01:07:31 +0000 (0:00:00.676)       0:04:37.092 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for services] *****************\nTuesday 13 January 2026  01:07:32 +0000 (0:00:00.833)       0:04:37.925 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf] **************************************\nTuesday 13 January 2026  01:07:35 +0000 (0:00:03.308)       0:04:41.234 ******* \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over Nova compute provider config] *******************\nTuesday 13 January 2026  01:07:41 +0000 (0:00:06.347)       0:04:47.582 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt configuration] **************************\nTuesday 13 January 2026  01:07:43 +0000 (0:00:01.393)       0:04:48.975 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\nTASK [nova-cell : Copying over libvirt TLS keys] *******************************\nTuesday 13 January 2026  01:07:46 +0000 (0:00:03.656)       0:04:52.631 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt SASL configuration] *********************\nTuesday 13 January 2026  01:07:47 +0000 (0:00:00.613)       0:04:53.245 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\nTASK [nova-cell : Copying files for nova-ssh] **********************************\nTuesday 13 January 2026  01:07:52 +0000 (0:00:05.086)       0:04:58.332 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [nova-cell : Copying VMware vCenter CA file] ******************************\nTuesday 13 January 2026  01:07:58 +0000 (0:00:06.504)       0:05:04.836 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying 'release' file for nova_compute] *********************\nTuesday 13 January 2026  01:07:59 +0000 (0:00:00.796)       0:05:05.633 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************\nTuesday 13 January 2026  01:08:00 +0000 (0:00:00.575)       0:05:06.209 ******* \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Copying over existing policy file] ***************************\nTuesday 13 January 2026  01:08:02 +0000 (0:00:02.045)       0:05:08.254 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over vendordata file to containers] ******************\nTuesday 13 January 2026  01:08:03 +0000 (0:00:01.211)       0:05:09.466 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check nova-cell containers] **********************************\nTuesday 13 January 2026  01:08:04 +0000 (0:00:00.834)       0:05:10.300 ******* \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 13 January 2026  01:08:07 +0000 (0:00:03.009)       0:05:13.309 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.745)       0:05:14.055 ******* \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.134)       0:05:14.189 ******* \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.130)       0:05:14.320 ******* \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.135)       0:05:14.456 ******* \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.137)       0:05:14.594 ******* \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 13 January 2026  01:08:08 +0000 (0:00:00.129)       0:05:14.723 ******* \n\nRUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************\nTuesday 13 January 2026  01:08:09 +0000 (0:00:00.362)       0:05:15.086 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************\nTuesday 13 January 2026  01:08:18 +0000 (0:00:09.493)       0:05:24.580 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************\nTuesday 13 January 2026  01:08:29 +0000 (0:00:10.888)       0:05:35.468 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************\nTuesday 13 January 2026  01:08:44 +0000 (0:00:15.133)       0:05:50.602 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************\nTuesday 13 January 2026  01:09:09 +0000 (0:00:24.319)       0:06:14.922 ******* \n\u001b[1;30mFAILED - RETRYING: [testbed-node-3]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************\nTuesday 13 January 2026  01:09:15 +0000 (0:00:06.168)       0:06:21.090 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-compute container] *******************\nTuesday 13 January 2026  01:09:15 +0000 (0:00:00.751)       0:06:21.842 ******* \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***\nTuesday 13 January 2026  01:09:38 +0000 (0:00:22.420)       0:06:44.262 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Waiting for nova-compute services to register themselves] ****\nTuesday 13 January 2026  01:09:38 +0000 (0:00:00.123)       0:06:44.386 ******* \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if nova-compute service failed to register] *************\nTuesday 13 January 2026  01:09:59 +0000 (0:00:20.923)       0:07:05.309 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Include discover_computes.yml] *******************************\nTuesday 13 January 2026  01:10:07 +0000 (0:00:07.722)       0:07:13.031 ******* \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 13 January 2026  01:10:10 +0000 (0:00:03.231)       0:07:16.263 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 13 January 2026  01:10:24 +0000 (0:00:13.846)       0:07:30.109 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if cell settings not found] *****************************\nTuesday 13 January 2026  01:10:25 +0000 (0:00:01.234)       0:07:31.344 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Discover nova hosts] *****************************************\nTuesday 13 January 2026  01:10:26 +0000 (0:00:01.284)       0:07:32.629 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************\nTuesday 13 January 2026  01:10:39 +0000 (0:00:12.681)       0:07:45.310 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Refresh nova scheduler cell cache] ***************************************\n\nTASK [nova : Refresh cell cache in nova scheduler] *****************************\nTuesday 13 January 2026  01:10:41 +0000 (0:00:01.689)       0:07:46.999 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova super conductor services] *****************************\n\nTASK [nova : Reload nova super conductor services to remove RPC version pin] ***\nTuesday 13 January 2026  01:10:42 +0000 (0:00:01.134)       0:07:48.134 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload Nova cell services] ***********************************************\n\nTASK [nova-cell : Reload nova cell services to remove RPC version cap] *********\nTuesday 13 January 2026  01:10:42 +0000 (0:00:00.530)       0:07:48.664 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova API services] *****************************************\n\nTASK [nova : Reload nova API services to remove RPC version pin] ***************\nTuesday 13 January 2026  01:10:44 +0000 (0:00:01.364)       0:07:50.029 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Run Nova API online data migrations] *************************************\n\nTASK [nova : Run Nova API online database migrations] **************************\nTuesday 13 January 2026  01:10:44 +0000 (0:00:00.747)       0:07:50.776 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Run Nova cell online data migrations] ************************************\n\nTASK [nova-cell : Run Nova cell online database migrations] ********************\nTuesday 13 January 2026  01:10:45 +0000 (0:00:00.622)       0:07:51.399 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=44  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:10:45 +0000 (0:00:00.409)       0:07:51.808 ******* \n=============================================================================== \nnova : Running Nova API bootstrap container ---------------------------- 31.02s\nnova-cell : Restart nova-libvirt container ----------------------------- 24.32s\nnova-cell : Restart nova-compute container ----------------------------- 22.42s\nnova-cell : Waiting for nova-compute services to register themselves --- 20.92s\nnova-cell : Running Nova cell bootstrap container ---------------------- 20.71s\nnova : Running Nova API bootstrap container ---------------------------- 18.25s\nnova : Create cell0 mappings ------------------------------------------- 15.50s\nnova : Restart nova-scheduler container -------------------------------- 15.33s\nnova-cell : Restart nova-ssh container --------------------------------- 15.13s\nnova-cell : Get a list of existing cells ------------------------------- 13.85s\nnova-cell : Create cell ------------------------------------------------ 13.77s\nnova-cell : Discover nova hosts ---------------------------------------- 12.68s\nnova-cell : Get a list of existing cells ------------------------------- 12.43s\nnova-cell : Get a list of existing cells ------------------------------- 11.76s\nnova : Restart nova-api container -------------------------------------- 11.10s\nnova-cell : Restart nova-novncproxy container -------------------------- 10.89s\nnova-cell : Restart nova-conductor container ---------------------------- 9.49s\nnova-cell : Fail if nova-compute service failed to register ------------- 7.72s\nservice-ks-register : nova | Granting user roles ------------------------ 7.28s\nservice-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 6.99s\n\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m\n\u001b[32m2026-01-13 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-01-13 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state SUCCESS\u001b[0m\n\u001b[32m2026-01-13 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:00.205)       0:00:00.205 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:00.246)       0:00:00.451 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m\n\nPLAY [Apply role octavia] ******************************************************\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:06:54 +0000 (0:00:00.382)       0:00:00.834 ******* \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : octavia | Creating services] ***********************\nTuesday 13 January 2026  01:06:55 +0000 (0:00:00.508)       0:00:01.343 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m\n\nTASK [service-ks-register : octavia | Creating endpoints] **********************\nTuesday 13 January 2026  01:06:58 +0000 (0:00:03.349)       0:00:04.692 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating projects] ***********************\nTuesday 13 January 2026  01:07:04 +0000 (0:00:06.076)       0:00:10.768 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating users] **************************\nTuesday 13 January 2026  01:07:07 +0000 (0:00:02.841)       0:00:13.609 ******* \n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating roles] **************************\nTuesday 13 January 2026  01:07:14 +0000 (0:00:06.613)       0:00:20.223 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : octavia | Granting user roles] *********************\nTuesday 13 January 2026  01:07:17 +0000 (0:00:02.887)       0:00:23.110 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\nTASK [octavia : Adding octavia related roles] **********************************\nTuesday 13 January 2026  01:07:22 +0000 (0:00:05.787)       0:00:28.898 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:07:36 +0000 (0:00:13.548)       0:00:42.447 ******* \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create amphora flavor] *****************************************\nTuesday 13 January 2026  01:07:37 +0000 (0:00:00.896)       0:00:43.344 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create nova keypair for amphora] *******************************\nTuesday 13 January 2026  01:07:42 +0000 (0:00:05.143)       0:00:48.488 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nTuesday 13 January 2026  01:07:46 +0000 (0:00:04.121)       0:00:52.609 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create security groups for octavia] ****************************\nTuesday 13 January 2026  01:07:49 +0000 (0:00:03.151)       0:00:55.761 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Add rules for security groups] *********************************\nTuesday 13 January 2026  01:07:58 +0000 (0:00:09.166)       0:01:04.927 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m\n\nTASK [octavia : Create loadbalancer management network] ************************\nTuesday 13 January 2026  01:08:13 +0000 (0:00:14.921)       0:01:19.849 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management subnet] *************************\nTuesday 13 January 2026  01:08:17 +0000 (0:00:03.755)       0:01:23.604 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management router for IPv6] ****************\nTuesday 13 January 2026  01:08:22 +0000 (0:00:05.177)       0:01:28.782 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update loadbalancer management subnet] *************************\nTuesday 13 January 2026  01:08:22 +0000 (0:00:00.224)       0:01:29.007 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:08:27 +0000 (0:00:04.178)       0:01:33.185 ******* \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create ports for Octavia health-manager nodes] *****************\nTuesday 13 January 2026  01:08:28 +0000 (0:00:00.941)       0:01:34.127 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Update Octavia health manager port host_id] ********************\nTuesday 13 January 2026  01:08:32 +0000 (0:00:04.274)       0:01:38.401 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Add Octavia port to openvswitch br-int] ************************\nTuesday 13 January 2026  01:08:36 +0000 (0:00:04.455)       0:01:42.856 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Install isc-dhcp-client package] *******************************\nTuesday 13 January 2026  01:08:37 +0000 (0:00:00.737)       0:01:43.594 ******* \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia dhclient conf] **********************************\nTuesday 13 January 2026  01:08:39 +0000 (0:00:02.051)       0:01:45.646 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia-interface service] ******************************\nTuesday 13 January 2026  01:08:40 +0000 (0:00:01.175)       0:01:46.821 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Restart octavia-interface.service if required] *****************\nTuesday 13 January 2026  01:08:41 +0000 (0:00:01.028)       0:01:47.849 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Enable and start octavia-interface.service] ********************\nTuesday 13 January 2026  01:08:43 +0000 (0:00:01.793)       0:01:49.643 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Wait for interface ohm0 ip appear] *****************************\nTuesday 13 January 2026  01:08:45 +0000 (0:00:01.707)       0:01:51.350 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Gather facts] **************************************************\nTuesday 13 January 2026  01:08:45 +0000 (0:00:00.609)       0:01:51.959 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:08:48 +0000 (0:00:02.448)       0:01:54.408 ******* \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Get amphora flavor info] ***************************************\nTuesday 13 January 2026  01:08:49 +0000 (0:00:00.765)       0:01:55.173 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nTuesday 13 January 2026  01:08:52 +0000 (0:00:03.287)       0:01:58.461 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get security groups for octavia] *******************************\nTuesday 13 January 2026  01:08:55 +0000 (0:00:03.250)       0:02:01.711 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Get loadbalancer management network] ***************************\nTuesday 13 January 2026  01:09:01 +0000 (0:00:05.803)       0:02:07.515 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia resources facts] ***********************************\nTuesday 13 January 2026  01:09:04 +0000 (0:00:02.819)       0:02:10.335 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Ensuring config directories exist] *****************************\nTuesday 13 January 2026  01:09:04 +0000 (0:00:00.299)       0:02:10.635 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Check if policies shall be overwritten] ************************\nTuesday 13 January 2026  01:09:06 +0000 (0:00:02.065)       0:02:12.700 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia policy file] ***************************************\nTuesday 13 January 2026  01:09:06 +0000 (0:00:00.139)       0:02:12.839 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over existing policy file] *****************************\nTuesday 13 January 2026  01:09:07 +0000 (0:00:00.481)       0:02:13.321 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:09:07 +0000 (0:00:00.691)       0:02:14.013 ******* \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over extra CA certificates] ********\nTuesday 13 January 2026  01:09:08 +0000 (0:00:00.543)       0:02:14.556 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***\nTuesday 13 January 2026  01:09:13 +0000 (0:00:04.758)       0:02:19.314 ******* \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****\nTuesday 13 January 2026  01:09:14 +0000 (0:00:00.955)       0:02:20.270 ******* \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over config.json files for services] *******************\nTuesday 13 January 2026  01:09:15 +0000 (0:00:00.844)       0:02:21.115 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over octavia-wsgi.conf] ********************************\nTuesday 13 January 2026  01:09:19 +0000 (0:00:04.640)       0:02:25.755 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\nTASK [octavia : Copying over octavia.conf] *************************************\nTuesday 13 January 2026  01:09:22 +0000 (0:00:02.388)       0:02:28.143 ******* \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over Octavia SSH key] **********************************\nTuesday 13 January 2026  01:09:38 +0000 (0:00:16.314)       0:02:44.457 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-worker] ******************\nTuesday 13 January 2026  01:09:39 +0000 (0:00:01.495)       0:02:45.953 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-housekeeping] ************\nTuesday 13 January 2026  01:09:46 +0000 (0:00:06.086)       0:02:52.039 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-health-manager] **********\nTuesday 13 January 2026  01:09:51 +0000 (0:00:05.063)       0:02:57.103 ******* \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Check octavia containers] **************************************\nTuesday 13 January 2026  01:09:55 +0000 (0:00:04.638)       0:03:01.741 ******* \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 13 January 2026  01:09:59 +0000 (0:00:03.506)       0:03:05.248 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Creating Octavia database] *************************************\nTuesday 13 January 2026  01:09:59 +0000 (0:00:00.297)       0:03:05.545 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database] *************************\nTuesday 13 January 2026  01:10:01 +0000 (0:00:01.830)       0:03:07.375 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia database user and setting permissions] ********\nTuesday 13 January 2026  01:10:03 +0000 (0:00:01.911)       0:03:09.286 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database user and setting permissions] ***\nTuesday 13 January 2026  01:10:05 +0000 (0:00:02.219)       0:03:11.506 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Running Octavia bootstrap container] ***************************\nTuesday 13 January 2026  01:10:08 +0000 (0:00:02.550)       0:03:14.057 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 13 January 2026  01:10:30 +0000 (0:00:22.676)       0:03:36.733 ******* \n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 13 January 2026  01:10:30 +0000 (0:00:00.064)       0:03:36.797 ******* \n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 13 January 2026  01:10:30 +0000 (0:00:00.068)       0:03:36.866 ******* \n\nRUNNING HANDLER [octavia : Restart octavia-api container] **********************\nTuesday 13 January 2026  01:10:30 +0000 (0:00:00.078)       0:03:36.944 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************\nTuesday 13 January 2026  01:10:46 +0000 (0:00:15.464)       0:03:52.409 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********\nTuesday 13 January 2026  01:10:52 +0000 (0:00:06.024)       0:03:58.433 ******* \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************\nTuesday 13 January 2026  01:11:02 +0000 (0:00:10.195)       0:04:08.629 ******* \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-worker container] *******************\nTuesday 13 January 2026  01:11:10 +0000 (0:00:08.123)       0:04:16.752 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:11:16 +0000 (0:00:05.519)       0:04:22.272 ******* \n=============================================================================== \noctavia : Running Octavia bootstrap container -------------------------- 22.68s\noctavia : Copying over octavia.conf ------------------------------------ 16.31s\noctavia : Restart octavia-api container -------------------------------- 15.46s\noctavia : Add rules for security groups -------------------------------- 14.92s\noctavia : Adding octavia related roles --------------------------------- 13.55s\noctavia : Restart octavia-health-manager container --------------------- 10.20s\noctavia : Create security groups for octavia ---------------------------- 9.17s\noctavia : Restart octavia-housekeeping container ------------------------ 8.12s\nservice-ks-register : octavia | Creating users -------------------------- 6.61s\noctavia : Copying certificate files for octavia-worker ------------------ 6.09s\nservice-ks-register : octavia | Creating endpoints ---------------------- 6.08s\noctavia : Restart octavia-driver-agent container ------------------------ 6.02s\noctavia : Get security groups for octavia ------------------------------- 5.80s\nservice-ks-register : octavia | Granting user roles --------------------- 5.79s\noctavia : Restart octavia-worker container ------------------------------ 5.52s\noctavia : Create loadbalancer management subnet ------------------------- 5.18s\noctavia : Create amphora flavor ----------------------------------------- 5.14s\noctavia : Copying certificate files for octavia-housekeeping ------------ 5.06s\nservice-cert-copy : octavia | Copying over extra CA certificates -------- 4.76s\noctavia : Copying over config.json files for services ------------------- 4.64s\n\u001b[32m2026-01-13 01:11:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:11:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-01-13 01:12:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\n\n--> DEPLOY IN A NUTSHELL -- END -- Tue Jan 13 01:12:19 UTC 2026",
                            "stdout_lines": [
                                "+ set -e",
                                "",
                                "# PULL IMAGES",
                                "",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "+ echo",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-01-13 00:38:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6e14501f-fcd9-40f9-abe9-92730c7570b3 (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6e14501f-fcd9-40f9-abe9-92730c7570b3 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2c21fb30-2dda-4d49-a0de-66a5504a8069 (wipe-partitions) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:38:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2c21fb30-2dda-4d49-a0de-66a5504a8069 (wipe-partitions) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wipe partitions] *********************************************************",
                                "",
                                "TASK [Find all logical devices owned by UID 167] *******************************",
                                "Tuesday 13 January 2026  00:39:00 +0000 (0:00:00.134)       0:00:00.134 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all rook related logical devices] *********************************",
                                "Tuesday 13 January 2026  00:39:01 +0000 (0:00:00.600)       0:00:00.735 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Find all logical devices with prefix ceph] *******************************",
                                "Tuesday 13 January 2026  00:39:01 +0000 (0:00:00.325)       0:00:01.060 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Remove all ceph related logical devices] *********************************",
                                "Tuesday 13 January 2026  00:39:02 +0000 (0:00:00.568)       0:00:01.629 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check device availability] ***********************************************",
                                "Tuesday 13 January 2026  00:39:02 +0000 (0:00:00.238)       0:00:01.867 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Wipe partitions with wipefs] *********************************************",
                                "Tuesday 13 January 2026  00:39:03 +0000 (0:00:01.160)       0:00:03.028 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Overwrite first 32M with zeros] ******************************************",
                                "Tuesday 13 January 2026  00:39:05 +0000 (0:00:01.475)       0:00:04.503 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Reload udev rules] *******************************************************",
                                "Tuesday 13 January 2026  00:39:09 +0000 (0:00:04.019)       0:00:08.523 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Request device events from the kernel] ***********************************",
                                "Tuesday 13 January 2026  00:39:09 +0000 (0:00:00.596)       0:00:09.120 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:39:10 +0000 (0:00:00.602)       0:00:09.722 ******* ",
                                "=============================================================================== ",
                                "Overwrite first 32M with zeros ------------------------------------------ 4.02s",
                                "Wipe partitions with wipefs --------------------------------------------- 1.48s",
                                "Check device availability ----------------------------------------------- 1.16s",
                                "Request device events from the kernel ----------------------------------- 0.60s",
                                "Find all logical devices owned by UID 167 ------------------------------- 0.60s",
                                "Reload udev rules ------------------------------------------------------- 0.60s",
                                "Find all logical devices with prefix ceph ------------------------------- 0.57s",
                                "Remove all rook related logical devices --------------------------------- 0.33s",
                                "Remove all ceph related logical devices --------------------------------- 0.24s",
                                "\u001b[32m2026-01-13 00:39:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f01a7713-5f5f-42c7-b15b-1d9d3ceefd85 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:39:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f01a7713-5f5f-42c7-b15b-1d9d3ceefd85 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  00:39:27 +0000 (0:00:00.254)       0:00:00.254 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 13 January 2026  00:39:28 +0000 (0:00:00.912)       0:00:01.166 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:39:29 +0000 (0:00:01.042)       0:00:02.209 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 13 January 2026  00:39:33 +0000 (0:00:04.791)       0:00:07.000 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:39:34 +0000 (0:00:00.495)       0:00:07.496 ******* ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.79s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.04s",
                                "osism.commons.facts : Create custom facts directory --------------------- 0.91s",
                                "Gather facts for all hosts ---------------------------------------------- 0.50s",
                                "\u001b[32m2026-01-13 00:39:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1aaf4672-483a-4030-a225-6812d7e6d7f1 (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:39:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1aaf4672-483a-4030-a225-6812d7e6d7f1 (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:39:41 +0000 (0:00:00.305)       0:00:00.305 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:39:41 +0000 (0:00:00.229)       0:00:00.535 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:41 +0000 (0:00:00.208)       0:00:00.743 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:42 +0000 (0:00:00.461)       0:00:01.205 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:42 +0000 (0:00:00.187)       0:00:01.393 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:42 +0000 (0:00:00.168)       0:00:01.561 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:42 +0000 (0:00:00.179)       0:00:01.741 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:42 +0000 (0:00:00.176)       0:00:01.917 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:43 +0000 (0:00:00.185)       0:00:02.102 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:43 +0000 (0:00:00.185)       0:00:02.288 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:43 +0000 (0:00:00.193)       0:00:02.482 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:43 +0000 (0:00:00.183)       0:00:02.666 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:44 +0000 (0:00:00.425)       0:00:03.091 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:44 +0000 (0:00:00.576)       0:00:03.668 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:45 +0000 (0:00:00.565)       0:00:04.234 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:46 +0000 (0:00:00.770)       0:00:05.005 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:46 +0000 (0:00:00.332)       0:00:05.337 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:46 +0000 (0:00:00.350)       0:00:05.688 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:46 +0000 (0:00:00.189)       0:00:05.877 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:47 +0000 (0:00:00.188)       0:00:06.066 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:47 +0000 (0:00:00.181)       0:00:06.247 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:47 +0000 (0:00:00.182)       0:00:06.429 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:47 +0000 (0:00:00.180)       0:00:06.610 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:47 +0000 (0:00:00.188)       0:00:06.798 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:48 +0000 (0:00:00.200)       0:00:06.998 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:48 +0000 (0:00:00.189)       0:00:07.188 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:49 +0000 (0:00:00.919)       0:00:08.107 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:49 +0000 (0:00:00.180)       0:00:08.288 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:49 +0000 (0:00:00.185)       0:00:08.473 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:49 +0000 (0:00:00.190)       0:00:08.663 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 13 January 2026  00:39:49 +0000 (0:00:00.192)       0:00:08.855 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.163)       0:00:09.018 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.121)       0:00:09.139 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.122)       0:00:09.262 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.112)       0:00:09.375 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.125)       0:00:09.500 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.153)       0:00:09.654 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 13 January 2026  00:39:50 +0000 (0:00:00.140)       0:00:09.795 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.308)       0:00:10.103 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.136)       0:00:10.240 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.129)       0:00:10.370 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.145)       0:00:10.515 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.133)       0:00:10.648 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.115)       0:00:10.764 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 13 January 2026  00:39:51 +0000 (0:00:00.136)       0:00:10.900 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 13 January 2026  00:39:52 +0000 (0:00:00.141)       0:00:11.041 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 13 January 2026  00:39:52 +0000 (0:00:00.134)       0:00:11.176 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 13 January 2026  00:39:52 +0000 (0:00:00.131)       0:00:11.307 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 13 January 2026  00:39:52 +0000 (0:00:00.124)       0:00:11.431 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-03961d85-1922-5669-8251-0ccc6cca9fac\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 13 January 2026  00:39:52 +0000 (0:00:00.375)       0:00:11.807 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:39:54 +0000 (0:00:01.731)       0:00:13.538 ******* ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:39:54 +0000 (0:00:00.250)       0:00:13.789 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:55 +0000 (0:00:00.222)       0:00:14.011 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:55 +0000 (0:00:00.345)       0:00:14.357 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:55 +0000 (0:00:00.172)       0:00:14.529 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:55 +0000 (0:00:00.182)       0:00:14.712 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:55 +0000 (0:00:00.186)       0:00:14.898 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:56 +0000 (0:00:00.184)       0:00:15.083 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:56 +0000 (0:00:00.508)       0:00:15.591 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:56 +0000 (0:00:00.180)       0:00:15.772 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:56 +0000 (0:00:00.184)       0:00:15.956 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:57 +0000 (0:00:00.188)       0:00:16.144 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:57 +0000 (0:00:00.392)       0:00:16.537 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:57 +0000 (0:00:00.390)       0:00:16.928 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:58 +0000 (0:00:00.404)       0:00:17.333 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:39:58 +0000 (0:00:00.395)       0:00:17.729 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:59 +0000 (0:00:00.322)       0:00:18.051 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:39:59 +0000 (0:00:00.354)       0:00:18.406 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:00 +0000 (0:00:00.669)       0:00:19.076 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:00 +0000 (0:00:00.191)       0:00:19.268 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:00 +0000 (0:00:00.218)       0:00:19.486 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:00 +0000 (0:00:00.201)       0:00:19.687 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:00 +0000 (0:00:00.195)       0:00:19.883 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:01 +0000 (0:00:00.208)       0:00:20.092 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:01 +0000 (0:00:00.190)       0:00:20.282 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:01 +0000 (0:00:00.187)       0:00:20.470 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:02 +0000 (0:00:00.686)       0:00:21.157 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:02 +0000 (0:00:00.156)       0:00:21.314 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:02 +0000 (0:00:00.166)       0:00:21.480 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:02 +0000 (0:00:00.177)       0:00:21.657 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.481)       0:00:22.139 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.110)       0:00:22.249 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.116)       0:00:22.365 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.103)       0:00:22.469 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.105)       0:00:22.575 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.106)       0:00:22.681 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.133)       0:00:22.815 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 13 January 2026  00:40:03 +0000 (0:00:00.116)       0:00:22.932 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.121)       0:00:23.054 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.114)       0:00:23.168 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.105)       0:00:23.274 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.102)       0:00:23.376 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.237)       0:00:23.614 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.136)       0:00:23.750 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 13 January 2026  00:40:04 +0000 (0:00:00.099)       0:00:23.849 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 13 January 2026  00:40:05 +0000 (0:00:00.142)       0:00:23.993 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 13 January 2026  00:40:05 +0000 (0:00:00.099)       0:00:24.092 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 13 January 2026  00:40:05 +0000 (0:00:00.109)       0:00:24.202 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 13 January 2026  00:40:05 +0000 (0:00:00.102)       0:00:24.304 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 13 January 2026  00:40:05 +0000 (0:00:00.180)       0:00:24.484 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:40:06 +0000 (0:00:00.995)       0:00:25.480 ******* ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:40:07 +0000 (0:00:00.482)       0:00:25.963 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:07 +0000 (0:00:00.194)       0:00:26.157 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:07 +0000 (0:00:00.307)       0:00:26.464 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:07 +0000 (0:00:00.175)       0:00:26.640 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:07 +0000 (0:00:00.176)       0:00:26.817 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.180)       0:00:26.997 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.170)       0:00:27.168 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.187)       0:00:27.355 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.166)       0:00:27.522 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.153)       0:00:27.676 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:08 +0000 (0:00:00.169)       0:00:27.846 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:09 +0000 (0:00:00.537)       0:00:28.383 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:09 +0000 (0:00:00.365)       0:00:28.749 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:10 +0000 (0:00:00.400)       0:00:29.149 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:40:10 +0000 (0:00:00.469)       0:00:29.618 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:10 +0000 (0:00:00.318)       0:00:29.937 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:11 +0000 (0:00:00.383)       0:00:30.320 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:11 +0000 (0:00:00.181)       0:00:30.502 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:11 +0000 (0:00:00.203)       0:00:30.706 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:11 +0000 (0:00:00.180)       0:00:30.886 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:12 +0000 (0:00:00.181)       0:00:31.068 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:12 +0000 (0:00:00.160)       0:00:31.228 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:12 +0000 (0:00:00.393)       0:00:31.621 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:12 +0000 (0:00:00.172)       0:00:31.794 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:12 +0000 (0:00:00.165)       0:00:31.959 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:13 +0000 (0:00:00.554)       0:00:32.514 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:13 +0000 (0:00:00.166)       0:00:32.681 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:13 +0000 (0:00:00.167)       0:00:32.848 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.173)       0:00:33.021 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.230)       0:00:33.252 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.191)       0:00:33.443 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.139)       0:00:33.582 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.131)       0:00:33.714 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 13 January 2026  00:40:14 +0000 (0:00:00.236)       0:00:33.950 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.109)       0:00:34.059 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.131)       0:00:34.190 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.121)       0:00:34.312 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.126)       0:00:34.438 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.136)       0:00:34.575 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.132)       0:00:34.707 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.109)       0:00:34.816 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 13 January 2026  00:40:15 +0000 (0:00:00.101)       0:00:34.918 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.131)       0:00:35.049 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.112)       0:00:35.162 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.120)       0:00:35.282 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.232)       0:00:35.514 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.100)       0:00:35.614 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.100)       0:00:35.715 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-e91d200a-cf56-55df-b2f8-08f15361112f\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 13 January 2026  00:40:16 +0000 (0:00:00.169)       0:00:35.885 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:40:17 +0000 (0:00:00.852)       0:00:36.738 ******* ",
                                "=============================================================================== ",
                                "Write configuration file ------------------------------------------------ 3.58s",
                                "Add known links to the list of available block devices ------------------ 1.11s",
                                "Add known partitions to the list of available block devices ------------- 1.09s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.96s",
                                "Add known partitions to the list of available block devices ------------- 0.92s",
                                "Add known links to the list of available block devices ------------------ 0.77s",
                                "Print configuration data ------------------------------------------------ 0.73s",
                                "Add known partitions to the list of available block devices ------------- 0.69s",
                                "Add known partitions to the list of available block devices ------------- 0.67s",
                                "Get initial list of available block devices ----------------------------- 0.63s",
                                "Add known links to the list of available block devices ------------------ 0.58s",
                                "Add known links to the list of available block devices ------------------ 0.57s",
                                "Generate lvm_volumes structure (block + wal) ---------------------------- 0.56s",
                                "Add known partitions to the list of available block devices ------------- 0.55s",
                                "Add known links to the list of available block devices ------------------ 0.54s",
                                "Add known links to the list of available block devices ------------------ 0.51s",
                                "Add known partitions to the list of available block devices ------------- 0.48s",
                                "Set DB devices config data ---------------------------------------------- 0.47s",
                                "Add known links to the list of available block devices ------------------ 0.47s",
                                "Print WAL devices ------------------------------------------------------- 0.47s",
                                "\u001b[32m2026-01-13 00:40:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5b6e9c71-ef81-48c9-b389-576a03ff85e9 (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-01-13 00:40:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-01-13 00:41:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "[master ea9e2af] 2026-01-13-00-41",
                                " 1 file changed, 30 insertions(+), 9 deletions(-)",
                                "\u001b[32m2026-01-13 00:41:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a91a25af-a18a-484c-9019-a75474dc5746 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:41:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a91a25af-a18a-484c-9019-a75474dc5746 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:41:12 +0000 (0:00:00.283)       0:00:00.283 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:41:12 +0000 (0:00:00.195)       0:00:00.479 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:12 +0000 (0:00:00.181)       0:00:00.661 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:13 +0000 (0:00:00.412)       0:00:01.073 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:13 +0000 (0:00:00.171)       0:00:01.245 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:13 +0000 (0:00:00.170)       0:00:01.415 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:13 +0000 (0:00:00.184)       0:00:01.600 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:14 +0000 (0:00:00.193)       0:00:01.794 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:14 +0000 (0:00:00.159)       0:00:01.953 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:14 +0000 (0:00:00.129)       0:00:02.083 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:14 +0000 (0:00:00.151)       0:00:02.234 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:14 +0000 (0:00:00.162)       0:00:02.396 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:15 +0000 (0:00:00.380)       0:00:02.777 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:15 +0000 (0:00:00.570)       0:00:03.347 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:16 +0000 (0:00:00.512)       0:00:03.860 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:16 +0000 (0:00:00.692)       0:00:04.553 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:17 +0000 (0:00:00.299)       0:00:04.852 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:17 +0000 (0:00:00.368)       0:00:05.221 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:17 +0000 (0:00:00.183)       0:00:05.405 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:17 +0000 (0:00:00.179)       0:00:05.584 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:18 +0000 (0:00:00.168)       0:00:05.753 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:18 +0000 (0:00:00.184)       0:00:05.937 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:18 +0000 (0:00:00.211)       0:00:06.148 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:18 +0000 (0:00:00.208)       0:00:06.357 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:18 +0000 (0:00:00.231)       0:00:06.589 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:19 +0000 (0:00:00.232)       0:00:06.821 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:20 +0000 (0:00:01.173)       0:00:07.995 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:20 +0000 (0:00:00.191)       0:00:08.186 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:20 +0000 (0:00:00.175)       0:00:08.362 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:20 +0000 (0:00:00.198)       0:00:08.560 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 13 January 2026  00:41:21 +0000 (0:00:00.202)       0:00:08.763 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 13 January 2026  00:41:21 +0000 (0:00:00.130)       0:00:08.893 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b9be54a9-cd9c-568c-9220-61b18da052d9'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '03961d85-1922-5669-8251-0ccc6cca9fac'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 13 January 2026  00:41:21 +0000 (0:00:00.175)       0:00:09.068 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 13 January 2026  00:41:23 +0000 (0:00:01.926)       0:00:10.995 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 13 January 2026  00:41:23 +0000 (0:00:00.145)       0:00:11.140 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 13 January 2026  00:41:24 +0000 (0:00:01.450)       0:00:12.591 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.162)       0:00:12.753 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.127)       0:00:12.880 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.364)       0:00:13.244 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.126)       0:00:13.371 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.135)       0:00:13.507 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 13 January 2026  00:41:25 +0000 (0:00:00.132)       0:00:13.639 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.162)       0:00:13.802 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.121)       0:00:13.923 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.134)       0:00:14.058 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.145)       0:00:14.203 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.154)       0:00:14.358 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.118)       0:00:14.476 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 13 January 2026  00:41:26 +0000 (0:00:00.125)       0:00:14.601 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 13 January 2026  00:41:27 +0000 (0:00:00.113)       0:00:14.715 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 13 January 2026  00:41:27 +0000 (0:00:00.296)       0:00:15.011 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 13 January 2026  00:41:27 +0000 (0:00:00.157)       0:00:15.169 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 13 January 2026  00:41:27 +0000 (0:00:00.159)       0:00:15.328 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 13 January 2026  00:41:28 +0000 (0:00:00.601)       0:00:15.930 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 13 January 2026  00:41:28 +0000 (0:00:00.509)       0:00:16.440 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.525)       0:00:16.965 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.127)       0:00:17.093 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.126)       0:00:17.219 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.122)       0:00:17.341 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.127)       0:00:17.469 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 13 January 2026  00:41:29 +0000 (0:00:00.133)       0:00:17.602 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 13 January 2026  00:41:30 +0000 (0:00:00.157)       0:00:17.760 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 13 January 2026  00:41:30 +0000 (0:00:00.462)       0:00:18.223 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 13 January 2026  00:41:30 +0000 (0:00:00.199)       0:00:18.422 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 13 January 2026  00:41:30 +0000 (0:00:00.161)       0:00:18.583 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.138)       0:00:18.722 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.152)       0:00:18.874 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.185)       0:00:19.060 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.142)       0:00:19.202 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.142)       0:00:19.345 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.157)       0:00:19.502 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 13 January 2026  00:41:31 +0000 (0:00:00.144)       0:00:19.646 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:41:32 +0000 (0:00:00.152)       0:00:19.799 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 13 January 2026  00:41:32 +0000 (0:00:00.158)       0:00:19.957 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 13 January 2026  00:41:32 +0000 (0:00:00.486)       0:00:20.443 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 13 January 2026  00:41:32 +0000 (0:00:00.172)       0:00:20.616 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.177)       0:00:20.794 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.173)       0:00:20.968 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.161)       0:00:21.129 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.161)       0:00:21.291 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.171)       0:00:21.462 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:41:33 +0000 (0:00:00.150)       0:00:21.613 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:41:34 +0000 (0:00:00.498)       0:00:22.111 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 13 January 2026  00:41:34 +0000 (0:00:00.544)       0:00:22.656 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 13 January 2026  00:41:35 +0000 (0:00:00.149)       0:00:22.806 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'vg_name': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'vg_name': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 13 January 2026  00:41:35 +0000 (0:00:00.166)       0:00:22.972 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 13 January 2026  00:41:35 +0000 (0:00:00.377)       0:00:23.349 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 13 January 2026  00:41:35 +0000 (0:00:00.167)       0:00:23.516 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 13 January 2026  00:41:35 +0000 (0:00:00.173)       0:00:23.690 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-03961d85-1922-5669-8251-0ccc6cca9fac\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-b9be54a9-cd9c-568c-9220-61b18da052d9\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-03961d85-1922-5669-8251-0ccc6cca9fac\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:41:36 +0000 (0:00:00.317)       0:00:24.007 ******* ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:41:36 +0000 (0:00:00.235)       0:00:24.242 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:36 +0000 (0:00:00.238)       0:00:24.480 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:37 +0000 (0:00:00.497)       0:00:24.978 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:37 +0000 (0:00:00.201)       0:00:25.179 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:37 +0000 (0:00:00.215)       0:00:25.395 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:38 +0000 (0:00:00.874)       0:00:26.270 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:38 +0000 (0:00:00.258)       0:00:26.528 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:39 +0000 (0:00:00.233)       0:00:26.762 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:39 +0000 (0:00:00.216)       0:00:26.978 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:39 +0000 (0:00:00.230)       0:00:27.209 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:39 +0000 (0:00:00.178)       0:00:27.387 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:40 +0000 (0:00:00.420)       0:00:27.808 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:40 +0000 (0:00:00.408)       0:00:28.217 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:40 +0000 (0:00:00.399)       0:00:28.616 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:41:41 +0000 (0:00:00.614)       0:00:29.231 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:42 +0000 (0:00:00.537)       0:00:29.769 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:42 +0000 (0:00:00.838)       0:00:30.607 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:43 +0000 (0:00:00.202)       0:00:30.810 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:43 +0000 (0:00:00.202)       0:00:31.013 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:43 +0000 (0:00:00.198)       0:00:31.211 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:43 +0000 (0:00:00.196)       0:00:31.408 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:43 +0000 (0:00:00.194)       0:00:31.603 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:44 +0000 (0:00:00.200)       0:00:31.804 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:44 +0000 (0:00:00.214)       0:00:32.019 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:44 +0000 (0:00:00.190)       0:00:32.209 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:45 +0000 (0:00:00.831)       0:00:33.040 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:45 +0000 (0:00:00.178)       0:00:33.219 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:46 +0000 (0:00:00.619)       0:00:33.838 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:41:46 +0000 (0:00:00.223)       0:00:34.062 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 13 January 2026  00:41:46 +0000 (0:00:00.206)       0:00:34.269 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 13 January 2026  00:41:46 +0000 (0:00:00.149)       0:00:34.418 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '2b3e8737-91e3-53c0-9b3a-5288a4111b63'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 13 January 2026  00:41:46 +0000 (0:00:00.199)       0:00:34.617 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 13 January 2026  00:41:48 +0000 (0:00:01.893)       0:00:36.511 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 13 January 2026  00:41:48 +0000 (0:00:00.149)       0:00:36.661 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 13 January 2026  00:41:50 +0000 (0:00:01.382)       0:00:38.043 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 13 January 2026  00:41:50 +0000 (0:00:00.157)       0:00:38.201 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 13 January 2026  00:41:50 +0000 (0:00:00.152)       0:00:38.354 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 13 January 2026  00:41:50 +0000 (0:00:00.144)       0:00:38.499 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 13 January 2026  00:41:50 +0000 (0:00:00.146)       0:00:38.645 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 13 January 2026  00:41:51 +0000 (0:00:00.368)       0:00:39.014 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 13 January 2026  00:41:51 +0000 (0:00:00.140)       0:00:39.155 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 13 January 2026  00:41:51 +0000 (0:00:00.161)       0:00:39.317 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 13 January 2026  00:41:51 +0000 (0:00:00.141)       0:00:39.458 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 13 January 2026  00:41:51 +0000 (0:00:00.166)       0:00:39.624 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.157)       0:00:39.782 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.155)       0:00:39.937 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.126)       0:00:40.064 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.142)       0:00:40.206 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.131)       0:00:40.338 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.133)       0:00:40.471 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 13 January 2026  00:41:52 +0000 (0:00:00.135)       0:00:40.607 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 13 January 2026  00:41:53 +0000 (0:00:00.348)       0:00:40.955 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 13 January 2026  00:41:53 +0000 (0:00:00.545)       0:00:41.500 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 13 January 2026  00:41:54 +0000 (0:00:00.556)       0:00:42.057 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 13 January 2026  00:41:54 +0000 (0:00:00.533)       0:00:42.590 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.160)       0:00:42.751 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.106)       0:00:42.858 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.103)       0:00:42.961 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.151)       0:00:43.112 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.132)       0:00:43.245 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.127)       0:00:43.373 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.127)       0:00:43.501 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 13 January 2026  00:41:55 +0000 (0:00:00.124)       0:00:43.625 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.335)       0:00:43.960 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.139)       0:00:44.100 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.128)       0:00:44.229 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.133)       0:00:44.362 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.133)       0:00:44.496 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:41:56 +0000 (0:00:00.143)       0:00:44.639 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.147)       0:00:44.787 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.148)       0:00:44.935 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.134)       0:00:45.069 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.148)       0:00:45.218 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.164)       0:00:45.383 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 13 January 2026  00:41:57 +0000 (0:00:00.145)       0:00:45.529 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 13 January 2026  00:41:58 +0000 (0:00:00.347)       0:00:45.876 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 13 January 2026  00:41:58 +0000 (0:00:00.158)       0:00:46.035 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 13 January 2026  00:41:58 +0000 (0:00:00.175)       0:00:46.211 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 13 January 2026  00:41:58 +0000 (0:00:00.154)       0:00:46.366 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 13 January 2026  00:41:58 +0000 (0:00:00.179)       0:00:46.545 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:41:59 +0000 (0:00:00.169)       0:00:46.715 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:41:59 +0000 (0:00:00.501)       0:00:47.216 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.515)       0:00:47.732 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.158)       0:00:47.890 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'vg_name': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'vg_name': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.195)       0:00:48.086 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.173)       0:00:48.259 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.176)       0:00:48.435 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 13 January 2026  00:42:00 +0000 (0:00:00.176)       0:00:48.612 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  00:42:01 +0000 (0:00:00.545)       0:00:49.157 ******* ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 13 January 2026  00:42:01 +0000 (0:00:00.259)       0:00:49.416 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:01 +0000 (0:00:00.250)       0:00:49.666 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:02 +0000 (0:00:00.429)       0:00:50.096 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:02 +0000 (0:00:00.236)       0:00:50.332 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:02 +0000 (0:00:00.176)       0:00:50.509 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:03 +0000 (0:00:00.192)       0:00:50.701 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:03 +0000 (0:00:00.209)       0:00:50.911 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:03 +0000 (0:00:00.592)       0:00:51.503 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:04 +0000 (0:00:00.200)       0:00:51.704 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:04 +0000 (0:00:00.222)       0:00:51.927 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:04 +0000 (0:00:00.193)       0:00:52.120 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:04 +0000 (0:00:00.393)       0:00:52.514 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:05 +0000 (0:00:00.406)       0:00:52.920 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:05 +0000 (0:00:00.446)       0:00:53.366 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 13 January 2026  00:42:06 +0000 (0:00:00.445)       0:00:53.811 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:06 +0000 (0:00:00.354)       0:00:54.166 ******* ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:06 +0000 (0:00:00.403)       0:00:54.569 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:07 +0000 (0:00:00.196)       0:00:54.765 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:07 +0000 (0:00:00.776)       0:00:55.542 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:08 +0000 (0:00:00.195)       0:00:55.738 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:08 +0000 (0:00:00.195)       0:00:55.933 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:08 +0000 (0:00:00.183)       0:00:56.117 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:08 +0000 (0:00:00.191)       0:00:56.308 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:08 +0000 (0:00:00.204)       0:00:56.512 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:09 +0000 (0:00:00.208)       0:00:56.721 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:09 +0000 (0:00:00.732)       0:00:57.453 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:09 +0000 (0:00:00.215)       0:00:57.669 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:10 +0000 (0:00:00.204)       0:00:57.874 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 13 January 2026  00:42:10 +0000 (0:00:00.186)       0:00:58.060 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 13 January 2026  00:42:10 +0000 (0:00:00.220)       0:00:58.280 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 13 January 2026  00:42:10 +0000 (0:00:00.291)       0:00:58.572 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e91d200a-cf56-55df-b2f8-08f15361112f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7ebda4f6-7b50-59b0-8273-b291dd7d1677'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 13 January 2026  00:42:11 +0000 (0:00:00.180)       0:00:58.753 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 13 January 2026  00:42:12 +0000 (0:00:01.810)       0:01:00.564 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 13 January 2026  00:42:13 +0000 (0:00:00.149)       0:01:00.713 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 13 January 2026  00:42:14 +0000 (0:00:01.301)       0:01:02.014 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 13 January 2026  00:42:14 +0000 (0:00:00.156)       0:01:02.171 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 13 January 2026  00:42:14 +0000 (0:00:00.125)       0:01:02.296 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 13 January 2026  00:42:14 +0000 (0:00:00.159)       0:01:02.456 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 13 January 2026  00:42:14 +0000 (0:00:00.130)       0:01:02.586 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 13 January 2026  00:42:15 +0000 (0:00:00.151)       0:01:02.738 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 13 January 2026  00:42:15 +0000 (0:00:00.142)       0:01:02.880 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 13 January 2026  00:42:15 +0000 (0:00:00.147)       0:01:03.028 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 13 January 2026  00:42:15 +0000 (0:00:00.363)       0:01:03.391 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 13 January 2026  00:42:15 +0000 (0:00:00.168)       0:01:03.560 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.166)       0:01:03.726 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.163)       0:01:03.890 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.141)       0:01:04.032 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.132)       0:01:04.164 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.157)       0:01:04.321 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.133)       0:01:04.454 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 13 January 2026  00:42:16 +0000 (0:00:00.155)       0:01:04.610 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 13 January 2026  00:42:17 +0000 (0:00:00.140)       0:01:04.750 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 13 January 2026  00:42:17 +0000 (0:00:00.512)       0:01:05.262 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 13 January 2026  00:42:18 +0000 (0:00:00.531)       0:01:05.794 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 13 January 2026  00:42:18 +0000 (0:00:00.735)       0:01:06.530 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 13 January 2026  00:42:18 +0000 (0:00:00.154)       0:01:06.684 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.118)       0:01:06.802 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.094)       0:01:06.897 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.123)       0:01:07.020 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.116)       0:01:07.136 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.129)       0:01:07.266 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.155)       0:01:07.422 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.131)       0:01:07.553 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 13 January 2026  00:42:19 +0000 (0:00:00.131)       0:01:07.684 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.124)       0:01:07.809 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.126)       0:01:07.936 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.338)       0:01:08.275 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.132)       0:01:08.407 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.132)       0:01:08.540 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 13 January 2026  00:42:20 +0000 (0:00:00.119)       0:01:08.660 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.136)       0:01:08.797 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.122)       0:01:08.919 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.119)       0:01:09.039 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.146)       0:01:09.186 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.146)       0:01:09.332 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.148)       0:01:09.480 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 13 January 2026  00:42:21 +0000 (0:00:00.143)       0:01:09.623 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 13 January 2026  00:42:22 +0000 (0:00:00.147)       0:01:09.771 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 13 January 2026  00:42:22 +0000 (0:00:00.358)       0:01:10.129 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 13 January 2026  00:42:22 +0000 (0:00:00.164)       0:01:10.294 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:42:22 +0000 (0:00:00.145)       0:01:10.439 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 13 January 2026  00:42:23 +0000 (0:00:00.525)       0:01:10.965 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 13 January 2026  00:42:23 +0000 (0:00:00.521)       0:01:11.487 ******* ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 13 January 2026  00:42:23 +0000 (0:00:00.133)       0:01:11.621 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'vg_name': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'vg_name': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 13 January 2026  00:42:24 +0000 (0:00:00.165)       0:01:11.787 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 13 January 2026  00:42:24 +0000 (0:00:00.148)       0:01:11.936 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 13 January 2026  00:42:24 +0000 (0:00:00.151)       0:01:12.087 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 13 January 2026  00:42:24 +0000 (0:00:00.165)       0:01:12.252 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-e91d200a-cf56-55df-b2f8-08f15361112f\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-e91d200a-cf56-55df-b2f8-08f15361112f\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:42:24 +0000 (0:00:00.140)       0:01:12.393 ******* ",
                                "=============================================================================== ",
                                "Create block VGs -------------------------------------------------------- 5.63s",
                                "Create block LVs -------------------------------------------------------- 4.13s",
                                "Gather DB+WAL VGs with total and available size in bytes ---------------- 1.79s",
                                "Gather DB VGs with total and available size in bytes -------------------- 1.66s",
                                "Add known partitions to the list of available block devices ------------- 1.61s",
                                "Gather WAL VGs with total and available size in bytes ------------------- 1.60s",
                                "Get list of Ceph PVs with associated VGs -------------------------------- 1.58s",
                                "Get list of Ceph LVs with associated VGs -------------------------------- 1.53s",
                                "Add known links to the list of available block devices ------------------ 1.34s",
                                "Add known partitions to the list of available block devices ------------- 1.17s",
                                "Print LVM report data --------------------------------------------------- 1.00s",
                                "Add known links to the list of available block devices ------------------ 0.87s",
                                "Add known partitions to the list of available block devices ------------- 0.83s",
                                "Create DB LVs for ceph_db_devices --------------------------------------- 0.80s",
                                "Add known partitions to the list of available block devices ------------- 0.78s",
                                "Print size needed for LVs on ceph_db_devices ---------------------------- 0.75s",
                                "Add known partitions to the list of available block devices ------------- 0.73s",
                                "Fail if block LV defined in lvm_volumes is missing ---------------------- 0.70s",
                                "Add known links to the list of available block devices ------------------ 0.69s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.69s",
                                "\u001b[32m2026-01-13 00:42:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f227694-2bc5-4186-89a0-6a2c85f9a271 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:42:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2f227694-2bc5-4186-89a0-6a2c85f9a271 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  00:42:41 +0000 (0:00:00.272)       0:00:00.272 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 13 January 2026  00:42:42 +0000 (0:00:01.049)       0:00:01.322 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  00:42:43 +0000 (0:00:01.226)       0:00:02.548 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 13 January 2026  00:42:48 +0000 (0:00:04.955)       0:00:07.503 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:42:49 +0000 (0:00:00.482)       0:00:07.985 ******* ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.96s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.23s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.05s",
                                "Gather facts for all hosts ---------------------------------------------- 0.48s",
                                "\u001b[32m2026-01-13 00:43:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask be4c30f5-fe4a-4450-a3bd-83c6374cf20b (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task be4c30f5-fe4a-4450-a3bd-83c6374cf20b (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Tuesday 13 January 2026  00:43:06 +0000 (0:00:00.265)       0:00:00.265 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Tuesday 13 January 2026  00:43:07 +0000 (0:00:00.253)       0:00:00.519 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Tuesday 13 January 2026  00:43:08 +0000 (0:00:01.301)       0:00:01.821 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Tuesday 13 January 2026  00:43:18 +0000 (0:00:10.520)       0:00:12.341 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Tuesday 13 January 2026  00:43:20 +0000 (0:00:01.017)       0:00:13.359 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Tuesday 13 January 2026  00:43:20 +0000 (0:00:00.973)       0:00:14.333 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Tuesday 13 January 2026  00:43:22 +0000 (0:00:01.200)       0:00:15.533 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Tuesday 13 January 2026  00:43:22 +0000 (0:00:00.163)       0:00:15.697 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Tuesday 13 January 2026  00:43:22 +0000 (0:00:00.163)       0:00:15.861 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Tuesday 13 January 2026  00:43:23 +0000 (0:00:00.982)       0:00:16.844 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Tuesday 13 January 2026  00:43:26 +0000 (0:00:03.331)       0:00:20.176 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.frr : Restart frr service] *********************",
                                "Tuesday 13 January 2026  00:43:28 +0000 (0:00:01.694)       0:00:21.870 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:43:30 +0000 (0:00:01.494)       0:00:23.365 ******* ",
                                "=============================================================================== ",
                                "osism.services.frr : Install frr package ------------------------------- 10.52s",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 3.33s",
                                "osism.services.frr : Manage frr service --------------------------------- 1.69s",
                                "osism.services.frr : Restart frr service -------------------------------- 1.49s",
                                "osism.services.frr : Pin frr package version ---------------------------- 1.30s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.20s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.02s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 0.98s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.97s",
                                "osism.services.frr : Include distribution specific install tasks -------- 0.25s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 0.16s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 0.16s",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- START -- Tue Jan 13 00:43:30 UTC 2026",
                                "",
                                "\u001b[32m2026-01-13 00:43:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:43:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "PLAY [Apply role geerlingguy.dotfiles] *****************************************",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****",
                                "Tuesday 13 January 2026  00:43:54 +0000 (0:00:00.446)       0:00:00.446 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********",
                                "Tuesday 13 January 2026  00:43:58 +0000 (0:00:03.847)       0:00:04.293 ******* ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***",
                                "Tuesday 13 January 2026  00:44:00 +0000 (0:00:02.188)       0:00:06.481 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.814918', 'end': '2026-01-13 00:43:58.820429', 'delta': '0:00:00.005511', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.803859', 'end': '2026-01-13 00:43:58.813355', 'delta': '0:00:00.009496', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:58.817015', 'end': '2026-01-13 00:43:58.826967', 'delta': '0:00:00.009952', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.124810', 'end': '2026-01-13 00:43:59.132657', 'delta': '0:00:00.007847', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.630323', 'end': '2026-01-13 00:43:59.637507', 'delta': '0:00:00.007184', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:43:59.371314', 'end': '2026-01-13 00:43:59.375075', 'delta': '0:00:00.003761', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-01-13 00:44:00.067080', 'end': '2026-01-13 00:44:00.075232', 'delta': '0:00:00.008152', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****",
                                "Tuesday 13 January 2026  00:44:02 +0000 (0:00:01.866)       0:00:08.348 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************",
                                "Tuesday 13 January 2026  00:44:04 +0000 (0:00:02.001)       0:00:10.349 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:44:07 +0000 (0:00:03.360)       0:00:13.710 ******* ",
                                "=============================================================================== ",
                                "geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 3.85s",
                                "geerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 3.36s",
                                "geerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.19s",
                                "geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 2.00s",
                                "geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 1.87s",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9d91f5a-e190-4cd8-a09b-07a90f2c0ae3 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de2d15df-73e5-4f2c-8364-14e2090e4924 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 18cb1beb-312f-4f92-8e89-583212497cc8 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:44:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77215c53-ff43-494e-9cab-86bfdd34cec3 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role homer] ********************************************************",
                                "",
                                "TASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***",
                                "Tuesday 13 January 2026  00:43:55 +0000 (0:00:01.241)       0:00:01.241 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create traefik external network] ******************",
                                "Tuesday 13 January 2026  00:43:55 +0000 (0:00:00.241)       0:00:01.482 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create required directories] **********************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:01.068)       0:00:02.551 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy config.yml configuration file] ***************",
                                "Tuesday 13 January 2026  00:43:58 +0000 (0:00:01.720)       0:00:04.272 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy docker-compose.yml file] *********************",
                                "Tuesday 13 January 2026  00:44:01 +0000 (0:00:02.720)       0:00:06.992 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Manage homer service] *****************************",
                                "Tuesday 13 January 2026  00:44:03 +0000 (0:00:01.731)       0:00:08.723 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.homer : Restart homer service] *****************",
                                "Tuesday 13 January 2026  00:44:31 +0000 (0:00:28.445)       0:00:37.169 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:44:35 +0000 (0:00:04.266)       0:00:41.435 ******* ",
                                "=============================================================================== ",
                                "osism.services.homer : Manage homer service ---------------------------- 28.45s",
                                "osism.services.homer : Restart homer service ---------------------------- 4.27s",
                                "osism.services.homer : Copy config.yml configuration file --------------- 2.72s",
                                "osism.services.homer : Copy docker-compose.yml file --------------------- 1.73s",
                                "osism.services.homer : Create required directories ---------------------- 1.72s",
                                "osism.services.homer : Create traefik external network ------------------ 1.07s",
                                "osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.24s",
                                "",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:00.581)       0:00:00.581 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:00.359)       0:00:00.940 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Tuesday 13 January 2026  00:43:59 +0000 (0:00:02.825)       0:00:03.766 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Tuesday 13 January 2026  00:44:02 +0000 (0:00:03.162)       0:00:06.929 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Tuesday 13 January 2026  00:44:36 +0000 (0:00:34.142)       0:00:41.071 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Tuesday 13 January 2026  00:44:38 +0000 (0:00:01.743)       0:00:42.815 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Tuesday 13 January 2026  00:44:39 +0000 (0:00:01.001)       0:00:43.816 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Tuesday 13 January 2026  00:44:42 +0000 (0:00:02.612)       0:00:46.429 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:02.059)       0:00:48.488 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:00.648)       0:00:49.137 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.413)       0:00:49.550 ******* ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Manage openstackclient service -------- 34.14s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 3.16s",
                                "osism.services.openstackclient : Create required directories ------------ 2.83s",
                                "osism.services.openstackclient : Restart openstackclient service -------- 2.61s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 2.06s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 1.74s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 1.00s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.65s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 0.41s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.36s",
                                "",
                                "",
                                "PLAY [Apply role phpmyadmin] ***************************************************",
                                "",
                                "TASK [osism.services.phpmyadmin : Create traefik external network] *************",
                                "Tuesday 13 January 2026  00:44:12 +0000 (0:00:00.354)       0:00:00.354 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Create required directories] *****************",
                                "Tuesday 13 January 2026  00:44:14 +0000 (0:00:01.132)       0:00:01.487 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************",
                                "Tuesday 13 January 2026  00:44:14 +0000 (0:00:00.602)       0:00:02.089 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************",
                                "Tuesday 13 January 2026  00:44:16 +0000 (0:00:01.619)       0:00:03.708 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******",
                                "Tuesday 13 January 2026  00:45:09 +0000 (0:00:53.611)       0:00:57.320 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:45:15 +0000 (0:00:05.575)       0:01:02.895 ******* ",
                                "=============================================================================== ",
                                "osism.services.phpmyadmin : Manage phpmyadmin service ------------------ 53.61s",
                                "osism.services.phpmyadmin : Restart phpmyadmin service ------------------ 5.58s",
                                "osism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.62s",
                                "osism.services.phpmyadmin : Create traefik external network ------------- 1.13s",
                                "osism.services.phpmyadmin : Create required directories ----------------- 0.60s",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1049fb69-c628-41aa-be00-1b600139bf4b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:43:55 +0000 (0:00:00.305)       0:00:00.305 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m",
                                "",
                                "PLAY [Apply role netdata] ******************************************************",
                                "",
                                "TASK [osism.services.netdata : Include distribution specific install tasks] ****",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:01.322)       0:00:01.628 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Remove old architecture-dependent repository] ***",
                                "Tuesday 13 January 2026  00:43:59 +0000 (0:00:03.084)       0:00:04.713 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install apt-transport-https package] ************",
                                "Tuesday 13 January 2026  00:44:02 +0000 (0:00:02.505)       0:00:07.218 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository gpg key] *************************",
                                "Tuesday 13 January 2026  00:44:05 +0000 (0:00:03.685)       0:00:10.904 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository] *********************************",
                                "Tuesday 13 January 2026  00:44:08 +0000 (0:00:02.058)       0:00:12.963 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install package netdata] ************************",
                                "Tuesday 13 January 2026  00:44:20 +0000 (0:00:12.305)       0:00:25.269 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include config tasks] ***************************",
                                "Tuesday 13 January 2026  00:45:04 +0000 (0:00:43.787)       0:01:09.057 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Copy configuration files] ***********************",
                                "Tuesday 13 January 2026  00:45:05 +0000 (0:00:01.388)       0:01:10.445 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***",
                                "Tuesday 13 January 2026  00:45:10 +0000 (0:00:04.818)       0:01:15.263 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Opt out from anonymous statistics] **************",
                                "Tuesday 13 January 2026  00:45:11 +0000 (0:00:00.968)       0:01:16.232 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add netdata user to docker group] ***************",
                                "Tuesday 13 January 2026  00:45:12 +0000 (0:00:01.259)       0:01:17.491 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Manage service netdata] *************************",
                                "Tuesday 13 January 2026  00:45:13 +0000 (0:00:01.343)       0:01:18.835 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include host type specific tasks] ***************",
                                "Tuesday 13 January 2026  00:45:16 +0000 (0:00:02.361)       0:01:21.197 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********",
                                "Tuesday 13 January 2026  00:45:18 +0000 (0:00:01.823)       0:01:23.020 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.netdata : Restart service netdata] *************",
                                "Tuesday 13 January 2026  00:45:19 +0000 (0:00:01.897)       0:01:24.918 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:45:23 +0000 (0:00:03.431)       0:01:28.349 ******* ",
                                "=============================================================================== ",
                                "osism.services.netdata : Install package netdata ----------------------- 43.79s",
                                "osism.services.netdata : Add repository -------------------------------- 12.31s",
                                "osism.services.netdata : Copy configuration files ----------------------- 4.82s",
                                "osism.services.netdata : Install apt-transport-https package ------------ 3.69s",
                                "osism.services.netdata : Restart service netdata ------------------------ 3.43s",
                                "osism.services.netdata : Include distribution specific install tasks ---- 3.08s",
                                "osism.services.netdata : Remove old architecture-dependent repository --- 2.51s",
                                "osism.services.netdata : Manage service netdata ------------------------- 2.36s",
                                "osism.services.netdata : Add repository gpg key ------------------------- 2.06s",
                                "osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.90s",
                                "osism.services.netdata : Include host type specific tasks --------------- 1.82s",
                                "osism.services.netdata : Include config tasks --------------------------- 1.39s",
                                "osism.services.netdata : Add netdata user to docker group --------------- 1.34s",
                                "Group hosts based on enabled services ----------------------------------- 1.32s",
                                "osism.services.netdata : Opt out from anonymous statistics -------------- 1.26s",
                                "osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 0.97s",
                                "\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:45:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab78807-a77c-43f5-9000-4296dc591a4d is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  00:43:47 +0000 (0:00:00.249)       0:00:00.249 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  00:43:49 +0000 (0:00:01.242)       0:00:01.491 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  00:43:53 +0000 (0:00:03.835)       0:00:05.326 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Tuesday 13 January 2026  00:43:54 +0000 (0:00:01.122)       0:00:06.449 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:43:58 +0000 (0:00:04.707)       0:00:11.157 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Tuesday 13 January 2026  00:44:00 +0000 (0:00:01.538)       0:00:12.696 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Tuesday 13 January 2026  00:44:03 +0000 (0:00:03.058)       0:00:15.754 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Tuesday 13 January 2026  00:44:04 +0000 (0:00:00.896)       0:00:16.650 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  00:44:05 +0000 (0:00:01.115)       0:00:17.765 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Tuesday 13 January 2026  00:44:11 +0000 (0:00:06.384)       0:00:24.149 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Tuesday 13 January 2026  00:44:14 +0000 (0:00:02.343)       0:00:26.493 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Tuesday 13 January 2026  00:44:15 +0000 (0:00:01.110)       0:00:27.603 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Tuesday 13 January 2026  00:44:16 +0000 (0:00:00.987)       0:00:28.591 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Tuesday 13 January 2026  00:44:17 +0000 (0:00:00.872)       0:00:29.464 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Tuesday 13 January 2026  00:44:20 +0000 (0:00:03.315)       0:00:32.779 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Tuesday 13 January 2026  00:44:23 +0000 (0:00:03.119)       0:00:35.898 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Tuesday 13 January 2026  00:44:27 +0000 (0:00:03.634)       0:00:39.533 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Tuesday 13 January 2026  00:44:31 +0000 (0:00:04.745)       0:00:44.278 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Tuesday 13 January 2026  00:44:34 +0000 (0:00:02.433)       0:00:46.711 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [common : Check common containers] ****************************************",
                                "Tuesday 13 January 2026  00:44:37 +0000 (0:00:02.608)       0:00:49.320 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Creating log volume] ********************************************",
                                "Tuesday 13 January 2026  00:44:40 +0000 (0:00:03.331)       0:00:52.651 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Link kolla_logs volume to /var/log/kolla] ***********************",
                                "Tuesday 13 January 2026  00:44:42 +0000 (0:00:02.177)       0:00:54.829 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:01.932)       0:00:56.761 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:00.077)       0:00:56.838 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:00.080)       0:00:56.919 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.413)       0:00:57.332 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.128)       0:00:57.461 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.121)       0:00:57.583 ******* ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.078)       0:00:57.661 ******* ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Tuesday 13 January 2026  00:44:45 +0000 (0:00:00.097)       0:00:57.759 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Tuesday 13 January 2026  00:45:16 +0000 (0:00:30.903)       0:01:28.662 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Tuesday 13 January 2026  00:45:50 +0000 (0:00:34.589)       0:02:03.252 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Tuesday 13 January 2026  00:45:53 +0000 (0:00:02.078)       0:02:05.331 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:46:02 +0000 (0:00:09.755)       0:02:15.086 ******* ",
                                "=============================================================================== ",
                                "common : Restart kolla-toolbox container ------------------------------- 34.59s",
                                "common : Restart fluentd container ------------------------------------- 30.90s",
                                "common : Restart cron container ----------------------------------------- 9.76s",
                                "common : Copying over config.json files for services -------------------- 6.38s",
                                "common : Ensuring config directories have correct owner and permission --- 4.75s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 4.71s",
                                "common : Ensuring config directories exist ------------------------------ 3.84s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 3.63s",
                                "common : Check common containers ---------------------------------------- 3.33s",
                                "common : Copying over fluentd.conf -------------------------------------- 3.32s",
                                "common : Copying over cron logrotate config file ------------------------ 3.12s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 3.06s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.61s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.43s",
                                "common : Find custom fluentd input config files ------------------------- 2.34s",
                                "common : Creating log volume -------------------------------------------- 2.18s",
                                "common : Initializing toolbox container using normal user --------------- 2.08s",
                                "common : Link kolla_logs volume to /var/log/kolla ----------------------- 1.93s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 1.54s",
                                "common : include_tasks -------------------------------------------------- 1.24s",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45ec02fd-74ee-40be-8c20-5bba8185c997 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.227)       0:00:00.227 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.270)       0:00:00.497 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.416)       0:00:00.914 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Tuesday 13 January 2026  00:46:10 +0000 (0:00:00.495)       0:00:01.409 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Tuesday 13 January 2026  00:46:10 +0000 (0:00:00.634)       0:00:02.043 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Check memcached container] ***********************************",
                                "Tuesday 13 January 2026  00:46:12 +0000 (0:00:01.907)       0:00:03.951 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Tuesday 13 January 2026  00:46:14 +0000 (0:00:01.570)       0:00:05.521 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:46:22 +0000 (0:00:08.193)       0:00:13.715 ******* ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container --------------------------------- 8.19s",
                                "memcached : Copying over config.json files for services ----------------- 1.91s",
                                "memcached : Check memcached container ----------------------------------- 1.57s",
                                "memcached : Ensuring config directories exist --------------------------- 0.63s",
                                "memcached : include_tasks ----------------------------------------------- 0.50s",
                                "Group hosts based on enabled services ----------------------------------- 0.42s",
                                "Group hosts based on Kolla action --------------------------------------- 0.27s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.216)       0:00:00.217 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.236)       0:00:00.453 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.396)       0:00:00.850 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.463)       0:00:01.314 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Tuesday 13 January 2026  00:46:10 +0000 (0:00:01.204)       0:00:02.519 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Tuesday 13 January 2026  00:46:13 +0000 (0:00:02.515)       0:00:05.034 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Check redis containers] ******************************************",
                                "Tuesday 13 January 2026  00:46:15 +0000 (0:00:02.372)       0:00:07.406 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 13 January 2026  00:46:17 +0000 (0:00:02.044)       0:00:09.451 ******* ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 13 January 2026  00:46:17 +0000 (0:00:00.109)       0:00:09.561 ******* ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 13 January 2026  00:46:18 +0000 (0:00:00.104)       0:00:09.666 ******* ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Tuesday 13 January 2026  00:46:18 +0000 (0:00:00.090)       0:00:09.756 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Tuesday 13 January 2026  00:46:27 +0000 (0:00:09.003)       0:00:18.759 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:46:37 +0000 (0:00:09.987)       0:00:28.747 ******* ",
                                "=============================================================================== ",
                                "redis : Restart redis-sentinel container -------------------------------- 9.99s",
                                "redis : Restart redis container ----------------------------------------- 9.00s",
                                "redis : Copying over default config.json files -------------------------- 2.52s",
                                "redis : Copying over redis config files --------------------------------- 2.37s",
                                "redis : Check redis containers ------------------------------------------ 2.05s",
                                "redis : Ensuring config directories exist ------------------------------- 1.20s",
                                "redis : include_tasks --------------------------------------------------- 0.46s",
                                "Group hosts based on enabled services ----------------------------------- 0.40s",
                                "redis : Flush handlers -------------------------------------------------- 0.30s",
                                "Group hosts based on Kolla action --------------------------------------- 0.24s",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9424a02-7da9-4373-8629-d432bfa169ab is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:46:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09713e6c-f6b7-47e5-9300-5209f889f05e is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.202)       0:00:00.202 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.690)       0:00:00.893 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.567)       0:00:01.461 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 13 January 2026  00:46:10 +0000 (0:00:01.174)       0:00:02.635 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 13 January 2026  00:46:12 +0000 (0:00:01.393)       0:00:04.029 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 13 January 2026  00:46:13 +0000 (0:00:01.466)       0:00:05.496 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Tuesday 13 January 2026  00:46:14 +0000 (0:00:01.101)       0:00:06.597 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Tuesday 13 January 2026  00:46:15 +0000 (0:00:00.770)       0:00:07.368 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Tuesday 13 January 2026  00:46:17 +0000 (0:00:01.851)       0:00:09.220 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Tuesday 13 January 2026  00:46:20 +0000 (0:00:02.935)       0:00:12.156 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Check openvswitch containers] ******************************",
                                "Tuesday 13 January 2026  00:46:21 +0000 (0:00:00.990)       0:00:13.146 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:24 +0000 (0:00:02.915)       0:00:16.062 ******* ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:24 +0000 (0:00:00.662)       0:00:16.724 ******* ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.266)       0:00:16.991 ******* ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.324)       0:00:17.315 ******* ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.356)       0:00:17.671 ******* ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 13 January 2026  00:46:26 +0000 (0:00:00.277)       0:00:17.948 ******* ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Tuesday 13 January 2026  00:46:26 +0000 (0:00:00.278)       0:00:18.227 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Tuesday 13 January 2026  00:46:37 +0000 (0:00:10.747)       0:00:28.974 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Tuesday 13 January 2026  00:46:38 +0000 (0:00:01.130)       0:00:30.104 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Tuesday 13 January 2026  00:46:47 +0000 (0:00:09.252)       0:00:39.357 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Tuesday 13 January 2026  00:46:55 +0000 (0:00:07.869)       0:00:47.226 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Tuesday 13 January 2026  00:46:58 +0000 (0:00:03.046)       0:00:50.273 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Tuesday 13 January 2026  00:47:02 +0000 (0:00:03.928)       0:00:54.201 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:47:10 +0000 (0:00:07.944)       0:01:02.146 ******* ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 17.20s",
                                "openvswitch : Restart openvswitch-db-server container ------------------ 10.75s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 7.87s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 3.93s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 3.05s",
                                "openvswitch : Copying over config.json files for services --------------- 2.94s",
                                "openvswitch : Check openvswitch containers ------------------------------ 2.92s",
                                "openvswitch : Flush Handlers -------------------------------------------- 2.16s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.85s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.47s",
                                "module-load : Load modules ---------------------------------------------- 1.39s",
                                "openvswitch : include_tasks --------------------------------------------- 1.17s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.13s",
                                "module-load : Drop module persistence ----------------------------------- 1.10s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.99s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 0.77s",
                                "Group hosts based on Kolla action --------------------------------------- 0.69s",
                                "Group hosts based on enabled services ----------------------------------- 0.57s",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ae35fb1d-5bc3-476d-a48a-9d1c5939422c is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Tuesday 13 January 2026  00:43:48 +0000 (0:00:00.170)       0:00:00.170 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Tuesday 13 January 2026  00:43:48 +0000 (0:00:00.686)       0:00:00.857 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Tuesday 13 January 2026  00:43:49 +0000 (0:00:00.508)       0:00:01.366 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Tuesday 13 January 2026  00:43:49 +0000 (0:00:00.591)       0:00:01.957 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Tuesday 13 January 2026  00:43:52 +0000 (0:00:02.678)       0:00:04.635 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Tuesday 13 January 2026  00:43:54 +0000 (0:00:01.815)       0:00:06.451 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:01.962)       0:00:08.414 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Tuesday 13 January 2026  00:43:57 +0000 (0:00:00.876)       0:00:09.290 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Tuesday 13 January 2026  00:43:58 +0000 (0:00:00.800)       0:00:10.090 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Tuesday 13 January 2026  00:43:58 +0000 (0:00:00.586)       0:00:10.677 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Tuesday 13 January 2026  00:43:59 +0000 (0:00:01.141)       0:00:11.818 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Tuesday 13 January 2026  00:44:00 +0000 (0:00:01.131)       0:00:12.950 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Tuesday 13 January 2026  00:44:06 +0000 (0:00:05.994)       0:00:18.944 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Tuesday 13 January 2026  00:44:08 +0000 (0:00:01.150)       0:00:20.095 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Tuesday 13 January 2026  00:44:10 +0000 (0:00:02.158)       0:00:22.253 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Tuesday 13 January 2026  00:44:11 +0000 (0:00:00.853)       0:00:23.106 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Tuesday 13 January 2026  00:44:12 +0000 (0:00:01.221)       0:00:24.327 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Tuesday 13 January 2026  00:44:13 +0000 (0:00:01.153)       0:00:25.481 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Tuesday 13 January 2026  00:44:15 +0000 (0:00:01.933)       0:00:27.414 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Tuesday 13 January 2026  00:44:16 +0000 (0:00:00.953)       0:00:28.367 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Tuesday 13 January 2026  00:44:17 +0000 (0:00:01.095)       0:00:29.463 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Tuesday 13 January 2026  00:44:18 +0000 (0:00:01.038)       0:00:30.502 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Tuesday 13 January 2026  00:44:19 +0000 (0:00:00.910)       0:00:31.413 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Tuesday 13 January 2026  00:44:19 +0000 (0:00:00.306)       0:00:31.719 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Tuesday 13 January 2026  00:44:20 +0000 (0:00:01.114)       0:00:32.833 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Tuesday 13 January 2026  00:44:22 +0000 (0:00:01.917)       0:00:34.751 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Tuesday 13 January 2026  00:44:23 +0000 (0:00:00.717)       0:00:35.469 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Tuesday 13 January 2026  00:44:26 +0000 (0:00:02.947)       0:00:38.417 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Tuesday 13 January 2026  00:44:26 +0000 (0:00:00.538)       0:00:38.955 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Tuesday 13 January 2026  00:44:28 +0000 (0:00:01.405)       0:00:40.361 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Tuesday 13 January 2026  00:44:30 +0000 (0:00:01.946)       0:00:42.308 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Tuesday 13 January 2026  00:44:30 +0000 (0:00:00.758)       0:00:43.066 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Tuesday 13 January 2026  00:44:31 +0000 (0:00:00.472)       0:00:43.539 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Tuesday 13 January 2026  00:44:33 +0000 (0:00:01.852)       0:00:45.391 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Tuesday 13 January 2026  00:44:36 +0000 (0:00:02.852)       0:00:48.244 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Tuesday 13 January 2026  00:44:36 +0000 (0:00:00.672)       0:00:48.916 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Tuesday 13 January 2026  00:45:20 +0000 (0:00:43.608)       0:01:32.525 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Tuesday 13 January 2026  00:45:20 +0000 (0:00:00.357)       0:01:32.883 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Tuesday 13 January 2026  00:45:21 +0000 (0:00:01.073)       0:01:33.957 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Tuesday 13 January 2026  00:45:23 +0000 (0:00:01.326)       0:01:35.284 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Tuesday 13 January 2026  00:45:48 +0000 (0:00:25.639)       0:02:00.923 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Tuesday 13 January 2026  00:45:49 +0000 (0:00:00.595)       0:02:01.519 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Tuesday 13 January 2026  00:45:50 +0000 (0:00:00.643)       0:02:02.162 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Tuesday 13 January 2026  00:45:50 +0000 (0:00:00.651)       0:02:02.813 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Tuesday 13 January 2026  00:45:51 +0000 (0:00:00.883)       0:02:03.697 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Tuesday 13 January 2026  00:45:51 +0000 (0:00:00.363)       0:02:04.060 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Tuesday 13 January 2026  00:45:52 +0000 (0:00:00.648)       0:02:04.708 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Tuesday 13 January 2026  00:45:53 +0000 (0:00:00.624)       0:02:05.333 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Tuesday 13 January 2026  00:45:54 +0000 (0:00:01.377)       0:02:06.711 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Tuesday 13 January 2026  00:45:55 +0000 (0:00:00.864)       0:02:07.576 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Tuesday 13 January 2026  00:45:55 +0000 (0:00:00.281)       0:02:07.857 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Tuesday 13 January 2026  00:45:56 +0000 (0:00:00.296)       0:02:08.153 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Tuesday 13 January 2026  00:45:56 +0000 (0:00:00.918)       0:02:09.072 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Tuesday 13 January 2026  00:45:57 +0000 (0:00:00.707)       0:02:09.779 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Tuesday 13 January 2026  00:46:00 +0000 (0:00:02.885)       0:02:12.665 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Tuesday 13 January 2026  00:46:01 +0000 (0:00:00.547)       0:02:13.213 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Tuesday 13 January 2026  00:46:01 +0000 (0:00:00.662)       0:02:13.875 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Tuesday 13 January 2026  00:46:02 +0000 (0:00:00.355)       0:02:14.231 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Tuesday 13 January 2026  00:46:02 +0000 (0:00:00.716)       0:02:14.947 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Tuesday 13 January 2026  00:46:03 +0000 (0:00:00.328)       0:02:15.276 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Tuesday 13 January 2026  00:46:03 +0000 (0:00:00.311)       0:02:15.587 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Tuesday 13 January 2026  00:46:03 +0000 (0:00:00.318)       0:02:15.906 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Tuesday 13 January 2026  00:46:04 +0000 (0:00:00.972)       0:02:16.878 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Tuesday 13 January 2026  00:46:05 +0000 (0:00:01.154)       0:02:18.032 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Tuesday 13 January 2026  00:46:07 +0000 (0:00:01.441)       0:02:19.474 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Tuesday 13 January 2026  00:46:19 +0000 (0:00:11.785)       0:02:31.260 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Tuesday 13 January 2026  00:46:19 +0000 (0:00:00.772)       0:02:32.033 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 13 January 2026  00:46:20 +0000 (0:00:00.517)       0:02:32.550 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 13 January 2026  00:46:20 +0000 (0:00:00.526)       0:02:33.077 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Tuesday 13 January 2026  00:46:21 +0000 (0:00:00.947)       0:02:34.024 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Tuesday 13 January 2026  00:46:22 +0000 (0:00:00.868)       0:02:34.892 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Tuesday 13 January 2026  00:46:24 +0000 (0:00:01.855)       0:02:36.748 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.915)       0:02:37.664 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.357)       0:02:38.022 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Tuesday 13 January 2026  00:46:26 +0000 (0:00:00.574)       0:02:38.596 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Tuesday 13 January 2026  00:46:26 +0000 (0:00:00.105)       0:02:38.702 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Tuesday 13 January 2026  00:46:26 +0000 (0:00:00.178)       0:02:38.880 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Tuesday 13 January 2026  00:46:27 +0000 (0:00:00.759)       0:02:39.639 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Tuesday 13 January 2026  00:46:28 +0000 (0:00:01.292)       0:02:40.932 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Tuesday 13 January 2026  00:46:29 +0000 (0:00:00.738)       0:02:41.670 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Tuesday 13 January 2026  00:46:29 +0000 (0:00:00.364)       0:02:42.035 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Tuesday 13 January 2026  00:46:36 +0000 (0:00:06.864)       0:02:48.899 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Tuesday 13 January 2026  00:46:48 +0000 (0:00:11.729)       0:03:00.628 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Tuesday 13 January 2026  00:46:49 +0000 (0:00:00.458)       0:03:01.087 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Tuesday 13 January 2026  00:46:49 +0000 (0:00:00.298)       0:03:01.386 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Tuesday 13 January 2026  00:46:49 +0000 (0:00:00.250)       0:03:01.636 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Tuesday 13 January 2026  00:46:50 +0000 (0:00:00.607)       0:03:02.243 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Tuesday 13 January 2026  00:46:51 +0000 (0:00:00.997)       0:03:03.241 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Tuesday 13 January 2026  00:46:51 +0000 (0:00:00.782)       0:03:04.023 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Tuesday 13 January 2026  00:46:52 +0000 (0:00:00.101)       0:03:04.125 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:00.985)       0:03:05.110 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:00.118)       0:03:05.228 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:00.116)       0:03:05.345 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:00.113)       0:03:05.459 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:00.129)       0:03:05.589 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Tuesday 13 January 2026  00:46:58 +0000 (0:00:05.335)       0:03:10.924 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Tuesday 13 January 2026  00:47:41 +0000 (0:00:42.203)       0:03:53.128 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Tuesday 13 January 2026  00:47:42 +0000 (0:00:01.205)       0:03:54.333 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Tuesday 13 January 2026  00:47:43 +0000 (0:00:01.557)       0:03:55.891 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Tuesday 13 January 2026  00:47:45 +0000 (0:00:01.209)       0:03:57.100 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Tuesday 13 January 2026  00:47:45 +0000 (0:00:00.101)       0:03:57.202 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Tuesday 13 January 2026  00:47:46 +0000 (0:00:01.545)       0:03:58.747 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.339)       0:03:59.086 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.952)       0:04:00.039 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Tuesday 13 January 2026  00:47:48 +0000 (0:00:00.131)       0:04:00.171 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Tuesday 13 January 2026  00:47:48 +0000 (0:00:00.195)       0:04:00.366 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Tuesday 13 January 2026  00:47:54 +0000 (0:00:06.380)       0:04:06.747 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Tuesday 13 January 2026  00:47:55 +0000 (0:00:01.215)       0:04:07.962 ******* ",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Tuesday 13 January 2026  00:48:07 +0000 (0:00:11.720)       0:04:19.682 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Tuesday 13 January 2026  00:48:08 +0000 (0:00:00.569)       0:04:20.252 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:48:08 +0000 (0:00:00.410)       0:04:20.662 ******* ",
                                "=============================================================================== ",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.61s",
                                "k3s_server_post : Wait for Cilium resources ---------------------------- 42.20s",
                                "k3s_server : Enable and check K3s service ------------------------------ 25.64s",
                                "k3s_agent : Manage k3s service ----------------------------------------- 11.79s",
                                "kubectl : Install required packages ------------------------------------ 11.73s",
                                "Manage labels ---------------------------------------------------------- 11.72s",
                                "kubectl : Add repository Debian ----------------------------------------- 6.86s",
                                "k9s : Install k9s packages ---------------------------------------------- 6.38s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 5.99s",
                                "k3s_server_post : Install Cilium ---------------------------------------- 5.34s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.95s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.89s",
                                "k3s_server : Detect Kubernetes version for label compatibility ---------- 2.85s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.68s",
                                "k3s_download : Download k3s binary armhf -------------------------------- 2.16s",
                                "k3s_prereq : Enable IPv6 router advertisements -------------------------- 1.96s",
                                "k3s_server : Copy vip manifest to first master -------------------------- 1.95s",
                                "k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 1.93s",
                                "k3s_server : Create custom resolv.conf for k3s -------------------------- 1.92s",
                                "Make kubeconfig available for use inside the manager service ------------ 1.86s",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15cdb02a-9808-43f4-bc82-148fef832dc1 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 33a1d39c-6cf7-41c4-b1e3-f05ff55a3212 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e398c6f6-1f7a-47da-aad9-f020dc0c55f2 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy kubeconfig to the configuration repository] *************************",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 13 January 2026  00:48:13 +0000 (0:00:00.176)       0:00:00.176 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 13 January 2026  00:48:14 +0000 (0:00:00.718)       0:00:00.895 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig file] ****************************",
                                "Tuesday 13 January 2026  00:48:15 +0000 (0:00:00.994)       0:00:01.890 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:48:15 +0000 (0:00:00.413)       0:00:02.303 ******* ",
                                "=============================================================================== ",
                                "Write kubeconfig file --------------------------------------------------- 0.99s",
                                "Get kubeconfig file ----------------------------------------------------- 0.72s",
                                "Change server address in the kubeconfig file ---------------------------- 0.41s",
                                "",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Tuesday 13 January 2026  00:48:13 +0000 (0:00:00.151)       0:00:00.151 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Tuesday 13 January 2026  00:48:13 +0000 (0:00:00.474)       0:00:00.625 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 13 January 2026  00:48:14 +0000 (0:00:00.776)       0:00:01.402 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 13 January 2026  00:48:15 +0000 (0:00:00.690)       0:00:02.093 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Tuesday 13 January 2026  00:48:16 +0000 (0:00:01.165)       0:00:03.258 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Tuesday 13 January 2026  00:48:17 +0000 (0:00:00.477)       0:00:03.735 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Tuesday 13 January 2026  00:48:18 +0000 (0:00:01.412)       0:00:05.148 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:00.756)       0:00:05.905 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:00.414)       0:00:06.319 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:00.221)       0:00:06.541 ******* ",
                                "=============================================================================== ",
                                "Make kubeconfig available for use inside the manager service ------------ 1.41s",
                                "Write kubeconfig file --------------------------------------------------- 1.17s",
                                "Create .kube directory -------------------------------------------------- 0.78s",
                                "Change server address in the kubeconfig inside the manager service ------ 0.76s",
                                "Get kubeconfig file ----------------------------------------------------- 0.69s",
                                "Change server address in the kubeconfig --------------------------------- 0.48s",
                                "Get home directory of operator user ------------------------------------- 0.47s",
                                "Set KUBECONFIG environment variable ------------------------------------- 0.41s",
                                "Enable kubectl command line completion ---------------------------------- 0.22s",
                                "",
                                "",
                                "PLAY [Set kolla_action_rabbitmq] ***********************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Tuesday 13 January 2026  00:46:30 +0000 (0:00:00.188)       0:00:00.188 ******* ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check RabbitMQ service] **************************************************",
                                "Tuesday 13 January 2026  00:46:30 +0000 (0:00:00.072)       0:00:00.261 ******* ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******",
                                "Tuesday 13 January 2026  00:46:33 +0000 (0:00:02.875)       0:00:03.136 ******* ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************",
                                "Tuesday 13 January 2026  00:46:33 +0000 (0:00:00.050)       0:00:03.186 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:46:33 +0000 (0:00:00.164)       0:00:03.351 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:46:33 +0000 (0:00:00.312)       0:00:03.664 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:46:34 +0000 (0:00:00.518)       0:00:04.183 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Tuesday 13 January 2026  00:46:34 +0000 (0:00:00.562)       0:00:04.746 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Tuesday 13 January 2026  00:46:35 +0000 (0:00:00.989)       0:00:05.736 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Tuesday 13 January 2026  00:46:35 +0000 (0:00:00.387)       0:00:06.123 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Tuesday 13 January 2026  00:46:36 +0000 (0:00:00.555)       0:00:06.679 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Tuesday 13 January 2026  00:46:37 +0000 (0:00:00.492)       0:00:07.172 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:46:37 +0000 (0:00:00.771)       0:00:07.943 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Tuesday 13 January 2026  00:46:38 +0000 (0:00:00.760)       0:00:08.704 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Tuesday 13 January 2026  00:46:39 +0000 (0:00:01.146)       0:00:09.850 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Tuesday 13 January 2026  00:46:40 +0000 (0:00:00.336)       0:00:10.186 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Tuesday 13 January 2026  00:46:40 +0000 (0:00:00.746)       0:00:10.933 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Tuesday 13 January 2026  00:46:41 +0000 (0:00:00.958)       0:00:11.891 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Tuesday 13 January 2026  00:46:43 +0000 (0:00:01.638)       0:00:13.529 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Tuesday 13 January 2026  00:46:45 +0000 (0:00:01.896)       0:00:15.426 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Tuesday 13 January 2026  00:46:47 +0000 (0:00:02.135)       0:00:17.562 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Tuesday 13 January 2026  00:46:49 +0000 (0:00:02.099)       0:00:19.662 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Tuesday 13 January 2026  00:46:51 +0000 (0:00:02.098)       0:00:21.760 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:01.632)       0:00:23.393 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:46:54 +0000 (0:00:01.696)       0:00:25.090 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check rabbitmq containers] ************************************",
                                "Tuesday 13 January 2026  00:46:55 +0000 (0:00:00.504)       0:00:25.594 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Creating rabbitmq volume] *************************************",
                                "Tuesday 13 January 2026  00:46:57 +0000 (0:00:02.301)       0:00:27.896 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Running RabbitMQ bootstrap container] *************************",
                                "Tuesday 13 January 2026  00:46:59 +0000 (0:00:01.363)       0:00:29.259 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Tuesday 13 January 2026  00:47:07 +0000 (0:00:08.407)       0:00:37.666 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 13 January 2026  00:47:07 +0000 (0:00:00.250)       0:00:37.917 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 13 January 2026  00:47:08 +0000 (0:00:00.571)       0:00:38.489 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 13 January 2026  00:47:08 +0000 (0:00:00.202)       0:00:38.691 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 13 January 2026  00:47:10 +0000 (0:00:01.622)       0:00:40.313 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 13 January 2026  00:48:02 +0000 (0:00:51.987)       0:01:32.301 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 13 January 2026  00:48:02 +0000 (0:00:00.818)       0:01:33.119 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 13 January 2026  00:48:03 +0000 (0:00:00.277)       0:01:33.397 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 13 January 2026  00:48:10 +0000 (0:00:07.097)       0:01:40.494 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 13 January 2026  00:48:20 +0000 (0:00:10.583)       0:01:51.078 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 13 January 2026  00:48:21 +0000 (0:00:00.724)       0:01:51.803 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 13 January 2026  00:48:21 +0000 (0:00:00.186)       0:01:51.989 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 13 January 2026  00:48:23 +0000 (0:00:01.741)       0:01:53.731 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Tuesday 13 January 2026  00:48:37 +0000 (0:00:14.303)       0:02:08.034 ******* ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:00.518)       0:02:08.552 ******* ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35menable_outward_rabbitmq_True\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35moutward_rabbitmq_restart\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq (outward)] *******************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq (outward) services] *************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq (outward) post-configuration] *****************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:48:40 +0000 (0:00:02.145)       0:02:10.698 ******* ",
                                "=============================================================================== ",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------- 76.87s",
                                "rabbitmq : Restart rabbitmq container ---------------------------------- 10.46s",
                                "rabbitmq : Running RabbitMQ bootstrap container ------------------------- 8.41s",
                                "Check RabbitMQ service -------------------------------------------------- 2.88s",
                                "rabbitmq : Check rabbitmq containers ------------------------------------ 2.30s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 2.15s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.14s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 2.12s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 2.10s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 2.10s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.90s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 1.70s",
                                "rabbitmq : Copying over config.json files for services ------------------ 1.64s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 1.63s",
                                "rabbitmq : Creating rabbitmq volume ------------------------------------- 1.36s",
                                "rabbitmq : Get container facts ------------------------------------------ 1.15s",
                                "rabbitmq : Get container facts ------------------------------------------ 0.99s",
                                "rabbitmq : Ensuring config directories exist ---------------------------- 0.96s",
                                "rabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 0.77s",
                                "rabbitmq : include_tasks ------------------------------------------------ 0.76s",
                                "\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ac9bf31-ed84-4bd5-a171-9bf739c5a717 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.147)       0:00:00.147 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:47:16 +0000 (0:00:00.820)       0:00:00.968 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.829)       0:00:01.797 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Tuesday 13 January 2026  00:47:18 +0000 (0:00:00.925)       0:00:02.723 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Tuesday 13 January 2026  00:47:19 +0000 (0:00:00.944)       0:00:03.667 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Tuesday 13 January 2026  00:47:20 +0000 (0:00:01.456)       0:00:05.124 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Tuesday 13 January 2026  00:47:21 +0000 (0:00:01.187)       0:00:06.311 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Check ovn-controller containers] ************************",
                                "Tuesday 13 January 2026  00:47:23 +0000 (0:00:02.037)       0:00:08.349 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Tuesday 13 January 2026  00:47:25 +0000 (0:00:01.860)       0:00:10.210 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Tuesday 13 January 2026  00:47:28 +0000 (0:00:02.695)       0:00:12.906 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:18.840)       0:00:31.746 ******* ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.050)       0:00:31.797 ******* ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.047)       0:00:31.844 ******* ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.053)       0:00:31.897 ******* ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.051)       0:00:31.949 ******* ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.049)       0:00:31.999 ******* ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Reload systemd config] ***********************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.049)       0:00:32.048 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Tuesday 13 January 2026  00:47:49 +0000 (0:00:02.017)       0:00:34.066 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  00:48:18 +0000 (0:00:28.701)       0:01:02.768 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:00.706)       0:01:03.474 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:00.410)       0:01:03.885 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Tuesday 13 January 2026  00:48:20 +0000 (0:00:00.941)       0:01:04.826 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Tuesday 13 January 2026  00:48:20 +0000 (0:00:00.290)       0:01:05.116 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Tuesday 13 January 2026  00:48:20 +0000 (0:00:00.265)       0:01:05.382 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Tuesday 13 January 2026  00:48:21 +0000 (0:00:00.299)       0:01:05.681 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Tuesday 13 January 2026  00:48:21 +0000 (0:00:00.381)       0:01:06.063 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Tuesday 13 January 2026  00:48:21 +0000 (0:00:00.253)       0:01:06.317 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:00.265)       0:01:06.582 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:00.249)       0:01:06.832 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:00.351)       0:01:07.184 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:00.254)       0:01:07.439 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Tuesday 13 January 2026  00:48:23 +0000 (0:00:00.231)       0:01:07.670 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Tuesday 13 January 2026  00:48:23 +0000 (0:00:00.257)       0:01:07.928 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Tuesday 13 January 2026  00:48:23 +0000 (0:00:00.369)       0:01:08.297 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Tuesday 13 January 2026  00:48:24 +0000 (0:00:00.326)       0:01:08.624 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Tuesday 13 January 2026  00:48:24 +0000 (0:00:00.352)       0:01:08.976 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Tuesday 13 January 2026  00:48:24 +0000 (0:00:00.301)       0:01:09.278 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  00:48:25 +0000 (0:00:00.343)       0:01:09.622 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************",
                                "Tuesday 13 January 2026  00:48:26 +0000 (0:00:00.913)       0:01:10.536 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************",
                                "Tuesday 13 January 2026  00:48:26 +0000 (0:00:00.466)       0:01:11.002 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check NB cluster status] ****************************************",
                                "Tuesday 13 January 2026  00:48:26 +0000 (0:00:00.452)       0:01:11.455 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check SB cluster status] ****************************************",
                                "Tuesday 13 January 2026  00:48:27 +0000 (0:00:00.586)       0:01:12.041 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***",
                                "Tuesday 13 January 2026  00:48:27 +0000 (0:00:00.337)       0:01:12.379 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***",
                                "Tuesday 13 January 2026  00:48:28 +0000 (0:00:00.330)       0:01:12.710 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************",
                                "Tuesday 13 January 2026  00:48:28 +0000 (0:00:00.351)       0:01:13.062 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************",
                                "Tuesday 13 January 2026  00:48:29 +0000 (0:00:00.568)       0:01:13.631 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  00:48:29 +0000 (0:00:00.299)       0:01:13.930 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  00:48:30 +0000 (0:00:01.425)       0:01:15.355 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Tuesday 13 January 2026  00:48:35 +0000 (0:00:04.484)       0:01:19.839 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:03.094)       0:01:22.934 ******* ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:00.069)       0:01:23.003 ******* ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:00.063)       0:01:23.067 ******* ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:00.067)       0:01:23.134 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Tuesday 13 January 2026  00:48:41 +0000 (0:00:02.377)       0:01:25.512 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Tuesday 13 January 2026  00:48:43 +0000 (0:00:02.393)       0:01:27.905 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Tuesday 13 January 2026  00:48:51 +0000 (0:00:07.858)       0:01:35.764 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Tuesday 13 January 2026  00:48:51 +0000 (0:00:00.131)       0:01:35.895 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Tuesday 13 January 2026  00:48:52 +0000 (0:00:00.709)       0:01:36.604 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Tuesday 13 January 2026  00:48:52 +0000 (0:00:00.524)       0:01:37.129 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Tuesday 13 January 2026  00:48:53 +0000 (0:00:00.687)       0:01:37.816 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Tuesday 13 January 2026  00:48:54 +0000 (0:00:00.828)       0:01:38.644 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Tuesday 13 January 2026  00:48:54 +0000 (0:00:00.792)       0:01:39.437 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Unset bootstrap args fact] **************************************",
                                "Tuesday 13 January 2026  00:48:55 +0000 (0:00:00.857)       0:01:40.295 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  00:48:56 +0000 (0:00:00.309)       0:01:40.604 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  00:48:57 +0000 (0:00:01.710)       0:01:42.314 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Tuesday 13 January 2026  00:49:02 +0000 (0:00:04.360)       0:01:46.675 ******* ",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:02.928)       0:01:49.604 ******* ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:00.060)       0:01:49.665 ******* ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:00.065)       0:01:49.731 ******* ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:00.063)       0:01:49.794 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Tuesday 13 January 2026  00:49:11 +0000 (0:00:06.304)       0:01:56.098 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Tuesday 13 January 2026  00:49:17 +0000 (0:00:06.211)       0:02:02.310 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Tuesday 13 January 2026  00:49:24 +0000 (0:00:06.433)       0:02:08.743 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Tuesday 13 January 2026  00:49:24 +0000 (0:00:00.144)       0:02:08.887 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Tuesday 13 January 2026  00:49:25 +0000 (0:00:00.893)       0:02:09.781 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Tuesday 13 January 2026  00:49:25 +0000 (0:00:00.663)       0:02:10.444 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Tuesday 13 January 2026  00:49:26 +0000 (0:00:00.788)       0:02:11.233 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Tuesday 13 January 2026  00:49:27 +0000 (0:00:00.612)       0:02:11.845 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Tuesday 13 January 2026  00:49:28 +0000 (0:00:00.787)       0:02:12.633 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:49:29 +0000 (0:00:00.910)       0:02:13.543 ******* ",
                                "=============================================================================== ",
                                "ovn-controller : Restart ovn-controller container ---------------------- 28.70s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 18.84s",
                                "ovn-db : Restart ovn-northd container ---------------------------------- 14.29s",
                                "ovn-db : Restart ovn-nb-db container ------------------------------------ 8.68s",
                                "ovn-db : Restart ovn-sb-db container ------------------------------------ 8.60s",
                                "ovn-db : Copying over config.json files for services -------------------- 4.48s",
                                "ovn-db : Copying over config.json files for services -------------------- 4.36s",
                                "ovn-db : Check ovn containers ------------------------------------------- 3.09s",
                                "ovn-db : Check ovn containers ------------------------------------------- 2.93s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.70s",
                                "ovn-controller : Copying over systemd override -------------------------- 2.04s",
                                "ovn-controller : Reload systemd config ---------------------------------- 2.02s",
                                "ovn-controller : Check ovn-controller containers ------------------------ 1.86s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.71s",
                                "ovn-controller : Copying over config.json files for services ------------ 1.46s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.43s",
                                "ovn-controller : Ensuring systemd override directory exists ------------- 1.19s",
                                "ovn-controller : Ensuring config directories exist ---------------------- 0.94s",
                                "ovn-db : Checking for any existing OVN DB container volumes ------------- 0.94s",
                                "ovn-controller : include_tasks ------------------------------------------ 0.93s",
                                "\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b89ee725-defa-4e6f-a85b-fcd8cc331623 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.246)       0:00:00.246 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:46:08 +0000 (0:00:00.276)       0:00:00.522 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.376)       0:00:00.898 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Tuesday 13 January 2026  00:46:09 +0000 (0:00:00.528)       0:00:01.427 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Tuesday 13 January 2026  00:46:10 +0000 (0:00:00.639)       0:00:02.066 ******* ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Tuesday 13 January 2026  00:46:11 +0000 (0:00:00.886)       0:00:02.952 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Tuesday 13 January 2026  00:46:11 +0000 (0:00:00.621)       0:00:03.574 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 13 January 2026  00:46:15 +0000 (0:00:03.190)       0:00:06.765 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 13 January 2026  00:46:15 +0000 (0:00:00.707)       0:00:07.472 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 13 January 2026  00:46:17 +0000 (0:00:01.591)       0:00:09.063 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Tuesday 13 January 2026  00:46:18 +0000 (0:00:00.923)       0:00:09.987 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Tuesday 13 January 2026  00:46:20 +0000 (0:00:02.258)       0:00:12.246 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Tuesday 13 January 2026  00:46:21 +0000 (0:00:01.046)       0:00:13.292 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Tuesday 13 January 2026  00:46:23 +0000 (0:00:02.216)       0:00:15.509 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:01.379)       0:00:16.889 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Tuesday 13 January 2026  00:46:27 +0000 (0:00:01.835)       0:00:18.724 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Tuesday 13 January 2026  00:46:29 +0000 (0:00:02.425)       0:00:21.150 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0', '__omit_place_holder__407d0674c1af89b0aa1d143b0a55ff06f98c36b0'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Tuesday 13 January 2026  00:46:32 +0000 (0:00:02.965)       0:00:24.115 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Tuesday 13 January 2026  00:46:35 +0000 (0:00:03.274)       0:00:27.390 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Tuesday 13 January 2026  00:46:38 +0000 (0:00:02.272)       0:00:29.663 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Tuesday 13 January 2026  00:46:42 +0000 (0:00:04.447)       0:00:34.111 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Tuesday 13 January 2026  00:46:43 +0000 (0:00:00.516)       0:00:34.628 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Tuesday 13 January 2026  00:46:45 +0000 (0:00:02.745)       0:00:37.373 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Tuesday 13 January 2026  00:46:48 +0000 (0:00:02.891)       0:00:40.264 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Tuesday 13 January 2026  00:46:50 +0000 (0:00:01.654)       0:00:41.919 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Tuesday 13 January 2026  00:46:52 +0000 (0:00:01.865)       0:00:43.785 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***",
                                "Tuesday 13 January 2026  00:46:53 +0000 (0:00:01.120)       0:00:44.905 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:46:57 +0000 (0:00:03.844)       0:00:48.749 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***",
                                "Tuesday 13 January 2026  00:46:57 +0000 (0:00:00.678)       0:00:49.428 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  00:46:58 +0000 (0:00:00.904)       0:00:50.332 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:47:00 +0000 (0:00:01.622)       0:00:51.954 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  00:47:00 +0000 (0:00:00.547)       0:00:52.502 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******",
                                "Tuesday 13 January 2026  00:47:01 +0000 (0:00:00.874)       0:00:53.376 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:47:03 +0000 (0:00:01.217)       0:00:54.594 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****",
                                "Tuesday 13 January 2026  00:47:04 +0000 (0:00:01.146)       0:00:55.740 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Tuesday 13 January 2026  00:47:04 +0000 (0:00:00.792)       0:00:56.533 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Tuesday 13 January 2026  00:47:06 +0000 (0:00:01.731)       0:00:58.265 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Tuesday 13 January 2026  00:47:08 +0000 (0:00:01.417)       0:00:59.682 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check loadbalancer containers] ****************************",
                                "Tuesday 13 January 2026  00:47:08 +0000 (0:00:00.809)       0:01:00.491 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Tuesday 13 January 2026  00:47:11 +0000 (0:00:02.483)       0:01:02.975 ******* ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Tuesday 13 January 2026  00:47:11 +0000 (0:00:00.523)       0:01:03.499 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:04.074)       0:01:07.573 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Tuesday 13 January 2026  00:47:16 +0000 (0:00:00.884)       0:01:08.458 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.954)       0:01:09.413 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Tuesday 13 January 2026  00:47:18 +0000 (0:00:01.121)       0:01:10.534 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Tuesday 13 January 2026  00:47:20 +0000 (0:00:01.960)       0:01:12.495 ******* ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Tuesday 13 January 2026  00:47:21 +0000 (0:00:00.762)       0:01:13.257 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:47:26 +0000 (0:00:04.488)       0:01:17.745 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Tuesday 13 January 2026  00:47:26 +0000 (0:00:00.730)       0:01:18.476 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Tuesday 13 January 2026  00:47:27 +0000 (0:00:01.082)       0:01:19.559 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Tuesday 13 January 2026  00:47:29 +0000 (0:00:01.439)       0:01:20.999 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Tuesday 13 January 2026  00:47:32 +0000 (0:00:02.584)       0:01:23.583 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Tuesday 13 January 2026  00:47:32 +0000 (0:00:00.322)       0:01:23.906 ******* ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Tuesday 13 January 2026  00:47:33 +0000 (0:00:00.849)       0:01:24.755 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:47:35 +0000 (0:00:02.571)       0:01:27.327 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Tuesday 13 January 2026  00:47:37 +0000 (0:00:01.563)       0:01:28.890 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Tuesday 13 January 2026  00:47:39 +0000 (0:00:02.151)       0:01:31.042 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Tuesday 13 January 2026  00:47:40 +0000 (0:00:00.876)       0:01:31.919 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Tuesday 13 January 2026  00:47:41 +0000 (0:00:01.256)       0:01:33.175 ******* ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Tuesday 13 January 2026  00:47:42 +0000 (0:00:00.762)       0:01:33.937 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:47:46 +0000 (0:00:04.160)       0:01:38.098 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:01.041)       0:01:39.139 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Tuesday 13 January 2026  00:47:48 +0000 (0:00:01.073)       0:01:40.212 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Tuesday 13 January 2026  00:47:49 +0000 (0:00:01.203)       0:01:41.416 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Tuesday 13 January 2026  00:47:52 +0000 (0:00:02.680)       0:01:44.097 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Tuesday 13 January 2026  00:47:53 +0000 (0:00:00.704)       0:01:44.801 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Tuesday 13 January 2026  00:47:53 +0000 (0:00:00.411)       0:01:45.213 ******* ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Tuesday 13 January 2026  00:47:54 +0000 (0:00:00.853)       0:01:46.067 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:00 +0000 (0:00:06.153)       0:01:52.220 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Tuesday 13 January 2026  00:48:01 +0000 (0:00:01.179)       0:01:53.400 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:48:03 +0000 (0:00:01.414)       0:01:54.815 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:48:05 +0000 (0:00:01.809)       0:01:56.624 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Tuesday 13 January 2026  00:48:06 +0000 (0:00:01.834)       0:01:58.459 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Tuesday 13 January 2026  00:48:07 +0000 (0:00:00.479)       0:01:58.938 ******* ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Tuesday 13 January 2026  00:48:07 +0000 (0:00:00.614)       0:01:59.553 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:12 +0000 (0:00:04.617)       0:02:04.170 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Tuesday 13 January 2026  00:48:16 +0000 (0:00:03.548)       0:02:07.719 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Tuesday 13 January 2026  00:48:19 +0000 (0:00:03.101)       0:02:10.820 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Tuesday 13 January 2026  00:48:20 +0000 (0:00:01.342)       0:02:12.163 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:01.898)       0:02:14.061 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Tuesday 13 January 2026  00:48:22 +0000 (0:00:00.431)       0:02:14.493 ******* ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Tuesday 13 January 2026  00:48:23 +0000 (0:00:00.793)       0:02:15.287 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:27 +0000 (0:00:03.351)       0:02:18.639 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Tuesday 13 January 2026  00:48:27 +0000 (0:00:00.627)       0:02:19.266 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:48:28 +0000 (0:00:00.644)       0:02:19.911 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:48:29 +0000 (0:00:01.361)       0:02:21.272 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Tuesday 13 January 2026  00:48:31 +0000 (0:00:02.179)       0:02:23.452 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Tuesday 13 January 2026  00:48:32 +0000 (0:00:00.744)       0:02:24.196 ******* ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Tuesday 13 January 2026  00:48:33 +0000 (0:00:00.896)       0:02:25.093 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:37 +0000 (0:00:04.091)       0:02:29.185 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Tuesday 13 January 2026  00:48:38 +0000 (0:00:01.228)       0:02:30.413 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:48:39 +0000 (0:00:01.063)       0:02:31.476 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:48:41 +0000 (0:00:01.194)       0:02:32.671 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Tuesday 13 January 2026  00:48:43 +0000 (0:00:01.929)       0:02:34.600 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Tuesday 13 January 2026  00:48:43 +0000 (0:00:00.298)       0:02:34.899 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Tuesday 13 January 2026  00:48:43 +0000 (0:00:00.577)       0:02:35.476 ******* ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Tuesday 13 January 2026  00:48:44 +0000 (0:00:00.921)       0:02:36.398 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:48 +0000 (0:00:03.384)       0:02:39.783 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Tuesday 13 January 2026  00:48:49 +0000 (0:00:00.908)       0:02:40.691 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Tuesday 13 January 2026  00:48:49 +0000 (0:00:00.795)       0:02:41.487 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Tuesday 13 January 2026  00:48:51 +0000 (0:00:01.286)       0:02:42.773 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Tuesday 13 January 2026  00:48:53 +0000 (0:00:02.094)       0:02:44.868 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Tuesday 13 January 2026  00:48:53 +0000 (0:00:00.628)       0:02:45.497 ******* ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Tuesday 13 January 2026  00:48:54 +0000 (0:00:00.947)       0:02:46.444 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:48:58 +0000 (0:00:03.601)       0:02:50.045 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Tuesday 13 January 2026  00:48:59 +0000 (0:00:01.195)       0:02:51.241 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Tuesday 13 January 2026  00:49:00 +0000 (0:00:00.915)       0:02:52.157 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Tuesday 13 January 2026  00:49:01 +0000 (0:00:01.372)       0:02:53.529 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Tuesday 13 January 2026  00:49:04 +0000 (0:00:02.467)       0:02:55.997 ******* ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:01.247)       0:02:57.244 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:49:09 +0000 (0:00:03.583)       0:03:00.827 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Tuesday 13 January 2026  00:49:09 +0000 (0:00:00.588)       0:03:01.416 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Tuesday 13 January 2026  00:49:10 +0000 (0:00:01.012)       0:03:02.428 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Tuesday 13 January 2026  00:49:12 +0000 (0:00:01.257)       0:03:03.685 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Tuesday 13 January 2026  00:49:14 +0000 (0:00:02.105)       0:03:05.791 ******* ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Tuesday 13 January 2026  00:49:15 +0000 (0:00:01.405)       0:03:07.197 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Tuesday 13 January 2026  00:49:18 +0000 (0:00:03.093)       0:03:10.290 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:49:20 +0000 (0:00:02.048)       0:03:12.338 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Tuesday 13 January 2026  00:49:23 +0000 (0:00:02.305)       0:03:14.644 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:49:26 +0000 (0:00:02.972)       0:03:17.616 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:49:27 +0000 (0:00:01.883)       0:03:19.500 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Tuesday 13 January 2026  00:49:29 +0000 (0:00:01.412)       0:03:20.912 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Tuesday 13 January 2026  00:49:29 +0000 (0:00:00.338)       0:03:21.251 ******* ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Tuesday 13 January 2026  00:49:31 +0000 (0:00:01.415)       0:03:22.667 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:49:32 +0000 (0:00:01.701)       0:03:24.368 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Tuesday 13 January 2026  00:49:33 +0000 (0:00:00.445)       0:03:24.814 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:49:34 +0000 (0:00:00.844)       0:03:25.659 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:49:34 +0000 (0:00:00.447)       0:03:26.106 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Tuesday 13 January 2026  00:49:35 +0000 (0:00:01.252)       0:03:27.358 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Tuesday 13 January 2026  00:49:36 +0000 (0:00:00.320)       0:03:27.679 ******* ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Tuesday 13 January 2026  00:49:37 +0000 (0:00:01.431)       0:03:29.110 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:49:41 +0000 (0:00:04.155)       0:03:33.265 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Tuesday 13 January 2026  00:49:43 +0000 (0:00:01.410)       0:03:34.676 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:49:45 +0000 (0:00:02.092)       0:03:36.769 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:49:46 +0000 (0:00:01.139)       0:03:37.908 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Tuesday 13 January 2026  00:49:48 +0000 (0:00:02.021)       0:03:39.930 ******* ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Tuesday 13 January 2026  00:49:49 +0000 (0:00:01.162)       0:03:41.092 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:49:52 +0000 (0:00:03.483)       0:03:44.576 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Tuesday 13 January 2026  00:49:53 +0000 (0:00:00.498)       0:03:45.075 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:49:54 +0000 (0:00:00.729)       0:03:45.805 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:49:55 +0000 (0:00:01.186)       0:03:46.991 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Tuesday 13 January 2026  00:49:57 +0000 (0:00:02.087)       0:03:49.078 ******* ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Tuesday 13 January 2026  00:49:59 +0000 (0:00:01.512)       0:03:50.590 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:50:03 +0000 (0:00:04.378)       0:03:54.969 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Tuesday 13 January 2026  00:50:04 +0000 (0:00:01.396)       0:03:56.366 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Tuesday 13 January 2026  00:50:05 +0000 (0:00:00.917)       0:03:57.283 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Tuesday 13 January 2026  00:50:07 +0000 (0:00:01.535)       0:03:58.819 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Tuesday 13 January 2026  00:50:09 +0000 (0:00:02.129)       0:04:00.949 ******* ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Tuesday 13 January 2026  00:50:11 +0000 (0:00:01.708)       0:04:02.657 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Tuesday 13 January 2026  00:50:11 +0000 (0:00:00.882)       0:04:03.539 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:50:16 +0000 (0:00:04.297)       0:04:07.837 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Tuesday 13 January 2026  00:50:17 +0000 (0:00:01.392)       0:04:09.230 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:50:19 +0000 (0:00:01.504)       0:04:10.734 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:50:21 +0000 (0:00:02.320)       0:04:13.054 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Tuesday 13 January 2026  00:50:24 +0000 (0:00:02.972)       0:04:16.027 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Tuesday 13 January 2026  00:50:25 +0000 (0:00:01.430)       0:04:17.458 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:50:27 +0000 (0:00:01.251)       0:04:18.709 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Tuesday 13 January 2026  00:50:28 +0000 (0:00:01.358)       0:04:20.068 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:50:30 +0000 (0:00:01.922)       0:04:21.990 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:50:32 +0000 (0:00:02.541)       0:04:24.531 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Tuesday 13 January 2026  00:50:36 +0000 (0:00:03.284)       0:04:27.816 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Tuesday 13 January 2026  00:50:37 +0000 (0:00:00.926)       0:04:28.743 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:50:38 +0000 (0:00:01.304)       0:04:30.047 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Tuesday 13 January 2026  00:50:39 +0000 (0:00:01.339)       0:04:31.387 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 13 January 2026  00:50:41 +0000 (0:00:01.591)       0:04:32.979 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 13 January 2026  00:50:44 +0000 (0:00:02.648)       0:04:35.627 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Tuesday 13 January 2026  00:50:47 +0000 (0:00:03.544)       0:04:39.172 ******* ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Tuesday 13 January 2026  00:50:49 +0000 (0:00:01.611)       0:04:40.783 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:50:52 +0000 (0:00:03.531)       0:04:44.315 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Tuesday 13 January 2026  00:50:53 +0000 (0:00:00.728)       0:04:45.043 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:50:55 +0000 (0:00:01.584)       0:04:46.627 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:50:56 +0000 (0:00:01.333)       0:04:47.960 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Tuesday 13 January 2026  00:50:58 +0000 (0:00:02.055)       0:04:50.016 ******* ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Tuesday 13 January 2026  00:50:59 +0000 (0:00:01.348)       0:04:51.365 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:51:05 +0000 (0:00:06.000)       0:04:57.366 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Tuesday 13 January 2026  00:51:06 +0000 (0:00:00.639)       0:04:58.005 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Tuesday 13 January 2026  00:51:07 +0000 (0:00:00.942)       0:04:58.948 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Tuesday 13 January 2026  00:51:08 +0000 (0:00:00.838)       0:04:59.787 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Tuesday 13 January 2026  00:51:09 +0000 (0:00:01.407)       0:05:01.194 ******* ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Tuesday 13 January 2026  00:51:10 +0000 (0:00:01.382)       0:05:02.576 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:51:15 +0000 (0:00:04.524)       0:05:07.101 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Tuesday 13 January 2026  00:51:16 +0000 (0:00:01.271)       0:05:08.372 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Tuesday 13 January 2026  00:51:17 +0000 (0:00:00.971)       0:05:09.344 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Tuesday 13 January 2026  00:51:18 +0000 (0:00:00.456)       0:05:09.801 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Tuesday 13 January 2026  00:51:19 +0000 (0:00:01.569)       0:05:11.370 ******* ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Tuesday 13 January 2026  00:51:21 +0000 (0:00:01.857)       0:05:13.228 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:51:24 +0000 (0:00:02.556)       0:05:15.785 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Tuesday 13 January 2026  00:51:24 +0000 (0:00:00.397)       0:05:16.183 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Tuesday 13 January 2026  00:51:25 +0000 (0:00:00.981)       0:05:17.164 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Tuesday 13 January 2026  00:51:26 +0000 (0:00:00.428)       0:05:17.593 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Tuesday 13 January 2026  00:51:27 +0000 (0:00:01.307)       0:05:18.901 ******* ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Tuesday 13 January 2026  00:51:29 +0000 (0:00:01.742)       0:05:20.644 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Tuesday 13 January 2026  00:51:35 +0000 (0:00:06.315)       0:05:26.960 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Tuesday 13 January 2026  00:51:35 +0000 (0:00:00.615)       0:05:27.575 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Tuesday 13 January 2026  00:51:37 +0000 (0:00:01.639)       0:05:29.215 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Tuesday 13 January 2026  00:51:38 +0000 (0:00:01.351)       0:05:30.566 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : swift] ****************************************************",
                                "Tuesday 13 January 2026  00:51:41 +0000 (0:00:02.165)       0:05:32.732 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Tuesday 13 January 2026  00:51:41 +0000 (0:00:00.323)       0:05:33.056 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Tuesday 13 January 2026  00:51:41 +0000 (0:00:00.311)       0:05:33.367 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Tuesday 13 January 2026  00:51:42 +0000 (0:00:00.637)       0:05:34.005 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Tuesday 13 January 2026  00:51:42 +0000 (0:00:00.328)       0:05:34.333 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Tuesday 13 January 2026  00:51:43 +0000 (0:00:00.312)       0:05:34.646 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Tuesday 13 January 2026  00:51:43 +0000 (0:00:00.844)       0:05:35.490 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Tuesday 13 January 2026  00:51:44 +0000 (0:00:00.699)       0:05:36.190 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Tuesday 13 January 2026  00:51:44 +0000 (0:00:00.359)       0:05:36.550 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Tuesday 13 January 2026  00:51:45 +0000 (0:00:00.895)       0:05:37.446 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Tuesday 13 January 2026  00:51:47 +0000 (0:00:01.336)       0:05:38.782 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Tuesday 13 January 2026  00:51:48 +0000 (0:00:00.960)       0:05:39.743 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Tuesday 13 January 2026  00:51:57 +0000 (0:00:09.590)       0:05:49.334 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Tuesday 13 January 2026  00:51:58 +0000 (0:00:00.816)       0:05:50.151 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Tuesday 13 January 2026  00:52:07 +0000 (0:00:08.751)       0:05:58.902 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Tuesday 13 January 2026  00:52:11 +0000 (0:00:04.183)       0:06:03.085 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Tuesday 13 January 2026  00:52:15 +0000 (0:00:04.425)       0:06:07.511 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Tuesday 13 January 2026  00:52:16 +0000 (0:00:00.367)       0:06:07.878 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Tuesday 13 January 2026  00:52:16 +0000 (0:00:00.356)       0:06:08.235 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Tuesday 13 January 2026  00:52:17 +0000 (0:00:00.736)       0:06:08.972 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Tuesday 13 January 2026  00:52:17 +0000 (0:00:00.334)       0:06:09.306 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Tuesday 13 January 2026  00:52:18 +0000 (0:00:00.379)       0:06:09.686 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Tuesday 13 January 2026  00:52:18 +0000 (0:00:00.339)       0:06:10.025 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:05.138)       0:06:15.163 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:52:24 +0000 (0:00:00.907)       0:06:16.071 ******* ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup haproxy container --------------------------- 9.59s",
                                "loadbalancer : Start backup proxysql container -------------------------- 8.75s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 6.32s",
                                "haproxy-config : Copying over designate haproxy config ------------------ 6.15s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 6.00s",
                                "loadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.14s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 4.62s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 4.52s",
                                "haproxy-config : Copying over barbican haproxy config ------------------- 4.49s",
                                "loadbalancer : Copying over proxysql config ----------------------------- 4.45s",
                                "loadbalancer : Start backup keepalived container ------------------------ 4.43s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 4.38s",
                                "haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.30s",
                                "loadbalancer : Wait for backup proxysql to start ------------------------ 4.18s",
                                "haproxy-config : Copying over cinder haproxy config --------------------- 4.16s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 4.16s",
                                "haproxy-config : Copying over horizon haproxy config -------------------- 4.09s",
                                "haproxy-config : Copying over aodh haproxy config ----------------------- 4.07s",
                                "service-cert-copy : loadbalancer | Copying over extra CA certificates --- 3.84s",
                                "haproxy-config : Copying over magnum haproxy config --------------------- 3.60s",
                                "\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:52:29 +0000 (0:00:00.287)       0:00:00.287 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:52:30 +0000 (0:00:00.286)       0:00:00.574 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  00:52:30 +0000 (0:00:00.456)       0:00:01.031 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Tuesday 13 January 2026  00:52:31 +0000 (0:00:00.475)       0:00:01.507 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Tuesday 13 January 2026  00:52:31 +0000 (0:00:00.707)       0:00:02.214 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  00:52:34 +0000 (0:00:02.228)       0:00:04.442 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Tuesday 13 January 2026  00:52:34 +0000 (0:00:00.574)       0:00:05.017 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:52:37 +0000 (0:00:03.053)       0:00:08.071 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Tuesday 13 January 2026  00:52:38 +0000 (0:00:00.996)       0:00:09.067 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Tuesday 13 January 2026  00:52:39 +0000 (0:00:00.818)       0:00:09.886 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Tuesday 13 January 2026  00:52:41 +0000 (0:00:02.177)       0:00:12.064 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Tuesday 13 January 2026  00:52:44 +0000 (0:00:03.351)       0:00:15.415 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Check opensearch containers] ********************************",
                                "Tuesday 13 January 2026  00:52:46 +0000 (0:00:01.693)       0:00:17.109 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  00:52:48 +0000 (0:00:02.152)       0:00:19.262 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.256)       0:00:19.519 ******* ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.058)       0:00:19.577 ******* ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.058)       0:00:19.636 ******* ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.060)       0:00:19.696 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.191)       0:00:19.888 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.457)       0:00:20.346 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Tuesday 13 January 2026  00:53:46 +0000 (0:00:56.663)       0:01:17.009 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  00:54:50 +0000 (0:01:03.539)       0:02:20.549 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Tuesday 13 January 2026  00:54:50 +0000 (0:00:00.709)       0:02:21.258 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch cluster to become healthy] **************",
                                "Tuesday 13 January 2026  00:54:53 +0000 (0:00:02.488)       0:02:23.746 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Tuesday 13 January 2026  00:54:55 +0000 (0:00:02.171)       0:02:25.918 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Tuesday 13 January 2026  00:54:58 +0000 (0:00:02.612)       0:02:28.531 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Tuesday 13 January 2026  00:55:01 +0000 (0:00:03.225)       0:02:31.756 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:55:03 +0000 (0:00:02.498)       0:02:34.255 ******* ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 63.54s",
                                "opensearch : Restart opensearch container ------------------------------ 56.66s",
                                "opensearch : Copying over opensearch service config file ---------------- 3.35s",
                                "opensearch : Create new log retention policy ---------------------------- 3.23s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 3.05s",
                                "opensearch : Check if a log retention policy exists --------------------- 2.61s",
                                "opensearch : Apply retention policy to existing indices ----------------- 2.50s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 2.49s",
                                "opensearch : Ensuring config directories exist -------------------------- 2.23s",
                                "opensearch : Copying over config.json files for services ---------------- 2.18s",
                                "opensearch : Wait for OpenSearch cluster to become healthy -------------- 2.17s",
                                "opensearch : Check opensearch containers -------------------------------- 2.15s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 1.69s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.00s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 0.82s",
                                "opensearch : include_tasks ---------------------------------------------- 0.71s",
                                "opensearch : Setting sysctl values -------------------------------------- 0.71s",
                                "opensearch : include_tasks ---------------------------------------------- 0.57s",
                                "opensearch : include_tasks ---------------------------------------------- 0.48s",
                                "opensearch : Perform a flush -------------------------------------------- 0.46s",
                                "\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 99df277e-6bfc-4fe3-b863-227d13b150e2 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 15d62ab2-5891-4568-a275-4a6f5a011aee is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Prepare deployment of Ceph services] *************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Tuesday 13 January 2026  00:43:48 +0000 (0:00:00.666)       0:00:00.666 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Tuesday 13 January 2026  00:43:49 +0000 (0:00:01.021)       0:00:01.687 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Tuesday 13 January 2026  00:43:51 +0000 (0:00:01.566)       0:00:03.254 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Tuesday 13 January 2026  00:43:51 +0000 (0:00:00.735)       0:00:03.990 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Tuesday 13 January 2026  00:43:52 +0000 (0:00:00.897)       0:00:04.887 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Tuesday 13 January 2026  00:43:53 +0000 (0:00:00.660)       0:00:05.548 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Tuesday 13 January 2026  00:43:53 +0000 (0:00:00.516)       0:00:06.065 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Tuesday 13 January 2026  00:43:54 +0000 (0:00:00.809)       0:00:06.874 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Tuesday 13 January 2026  00:43:55 +0000 (0:00:00.680)       0:00:07.555 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:00.641)       0:00:08.197 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Tuesday 13 January 2026  00:43:56 +0000 (0:00:00.642)       0:00:08.839 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Tuesday 13 January 2026  00:43:57 +0000 (0:00:01.208)       0:00:10.048 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Tuesday 13 January 2026  00:44:01 +0000 (0:00:03.342)       0:00:13.391 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Tuesday 13 January 2026  00:44:02 +0000 (0:00:00.772)       0:00:14.163 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Tuesday 13 January 2026  00:44:03 +0000 (0:00:01.044)       0:00:15.208 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Tuesday 13 January 2026  00:44:03 +0000 (0:00:00.405)       0:00:15.614 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-01-13 00:43:58.698129', 'end': '2026-01-13 00:43:59.016841', 'delta': '0:00:00.318712', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-01-13 00:43:59.859242', 'end': '2026-01-13 00:44:00.158616', 'delta': '0:00:00.299374', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-01-13 00:44:00.795375', 'end': '2026-01-13 00:44:01.103968', 'delta': '0:00:00.308593', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Tuesday 13 January 2026  00:44:03 +0000 (0:00:00.285)       0:00:15.899 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Tuesday 13 January 2026  00:44:05 +0000 (0:00:01.664)       0:00:17.564 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Tuesday 13 January 2026  00:44:06 +0000 (0:00:00.757)       0:00:18.322 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Tuesday 13 January 2026  00:44:08 +0000 (0:00:02.089)       0:00:20.412 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 13 January 2026  00:44:10 +0000 (0:00:01.944)       0:00:22.356 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Tuesday 13 January 2026  00:44:11 +0000 (0:00:01.503)       0:00:23.859 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Tuesday 13 January 2026  00:44:11 +0000 (0:00:00.126)       0:00:23.986 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 13 January 2026  00:44:12 +0000 (0:00:00.377)       0:00:24.364 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Tuesday 13 January 2026  00:44:13 +0000 (0:00:01.133)       0:00:25.497 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Tuesday 13 January 2026  00:44:14 +0000 (0:00:01.175)       0:00:26.672 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Tuesday 13 January 2026  00:44:15 +0000 (0:00:00.680)       0:00:27.352 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Tuesday 13 January 2026  00:44:16 +0000 (0:00:01.590)       0:00:28.943 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Tuesday 13 January 2026  00:44:17 +0000 (0:00:00.743)       0:00:29.686 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Tuesday 13 January 2026  00:44:18 +0000 (0:00:00.825)       0:00:30.512 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Tuesday 13 January 2026  00:44:19 +0000 (0:00:00.678)       0:00:31.190 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-02-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-00-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Tuesday 13 January 2026  00:44:20 +0000 (0:00:01.230)       0:00:32.421 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16', 'scsi-SQEMU_QEMU_HARDDISK_4dcaa69c-5414-4861-9f75-cc0da42200e7-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16', 'scsi-SQEMU_QEMU_HARDDISK_83689365-4423-433a-82c0-63cbcaedfdf8-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-02-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16', 'scsi-SQEMU_QEMU_HARDDISK_d36bd727-f6fd-4e09-af6c-5d1752a9fb11-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-00-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Tuesday 13 January 2026  00:44:21 +0000 (0:00:01.434)       0:00:33.855 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Tuesday 13 January 2026  00:44:23 +0000 (0:00:01.356)       0:00:35.212 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 13 January 2026  00:44:23 +0000 (0:00:00.548)       0:00:35.761 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 13 January 2026  00:44:24 +0000 (0:00:01.247)       0:00:37.009 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 13 January 2026  00:44:25 +0000 (0:00:00.834)       0:00:37.843 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 13 January 2026  00:44:26 +0000 (0:00:00.916)       0:00:38.760 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Tuesday 13 January 2026  00:44:27 +0000 (0:00:00.852)       0:00:39.613 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Tuesday 13 January 2026  00:44:32 +0000 (0:00:05.016)       0:00:44.630 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Tuesday 13 January 2026  00:44:33 +0000 (0:00:01.108)       0:00:45.738 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 13 January 2026  00:44:34 +0000 (0:00:01.119)       0:00:46.858 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 13 January 2026  00:44:35 +0000 (0:00:00.611)       0:00:47.469 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 13 January 2026  00:44:36 +0000 (0:00:00.647)       0:00:48.117 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 13 January 2026  00:44:36 +0000 (0:00:00.596)       0:00:48.713 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 13 January 2026  00:44:37 +0000 (0:00:00.847)       0:00:49.561 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 13 January 2026  00:44:37 +0000 (0:00:00.328)       0:00:49.889 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 13 January 2026  00:44:38 +0000 (0:00:00.349)       0:00:50.239 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 13 January 2026  00:44:38 +0000 (0:00:00.477)       0:00:50.716 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 13 January 2026  00:44:39 +0000 (0:00:00.382)       0:00:51.099 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Tuesday 13 January 2026  00:44:39 +0000 (0:00:00.883)       0:00:51.982 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Tuesday 13 January 2026  00:44:40 +0000 (0:00:00.984)       0:00:52.967 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:44:43 +0000 (0:00:02.195)       0:00:55.162 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:44:44 +0000 (0:00:01.695)       0:00:56.858 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:44:46 +0000 (0:00:01.648)       0:00:58.506 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:44:48 +0000 (0:00:01.632)       0:01:00.139 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:44:49 +0000 (0:00:01.005)       0:01:01.144 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:44:50 +0000 (0:00:01.056)       0:01:02.200 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:44:50 +0000 (0:00:00.852)       0:01:03.053 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:44:52 +0000 (0:00:01.470)       0:01:04.523 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:44:52 +0000 (0:00:00.544)       0:01:05.067 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:44:53 +0000 (0:00:00.802)       0:01:05.870 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:44:54 +0000 (0:00:01.099)       0:01:06.970 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:44:56 +0000 (0:00:01.579)       0:01:08.549 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:44:57 +0000 (0:00:00.635)       0:01:09.185 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:44:58 +0000 (0:00:01.001)       0:01:10.186 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:44:58 +0000 (0:00:00.630)       0:01:10.817 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:44:59 +0000 (0:00:00.921)       0:01:11.739 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:45:00 +0000 (0:00:00.679)       0:01:12.418 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:45:01 +0000 (0:00:01.113)       0:01:13.532 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:45:02 +0000 (0:00:00.854)       0:01:14.386 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:45:03 +0000 (0:00:01.465)       0:01:15.851 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:45:04 +0000 (0:00:00.649)       0:01:16.500 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Tuesday 13 January 2026  00:45:05 +0000 (0:00:01.315)       0:01:17.816 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Tuesday 13 January 2026  00:45:07 +0000 (0:00:01.946)       0:01:19.762 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Tuesday 13 January 2026  00:45:10 +0000 (0:00:03.006)       0:01:22.768 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Tuesday 13 January 2026  00:45:11 +0000 (0:00:01.013)       0:01:23.781 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Tuesday 13 January 2026  00:45:12 +0000 (0:00:00.648)       0:01:24.430 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Tuesday 13 January 2026  00:45:13 +0000 (0:00:00.779)       0:01:25.210 ******* ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Tuesday 13 January 2026  00:45:14 +0000 (0:00:01.510)       0:01:26.720 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Tuesday 13 January 2026  00:45:16 +0000 (0:00:02.169)       0:01:28.889 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Tuesday 13 January 2026  00:45:17 +0000 (0:00:01.037)       0:01:29.927 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Tuesday 13 January 2026  00:45:18 +0000 (0:00:00.968)       0:01:30.895 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Tuesday 13 January 2026  00:45:19 +0000 (0:00:00.561)       0:01:31.457 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Tuesday 13 January 2026  00:45:20 +0000 (0:00:01.058)       0:01:32.516 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Tuesday 13 January 2026  00:46:21 +0000 (0:01:00.887)       0:02:33.404 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Tuesday 13 January 2026  00:46:22 +0000 (0:00:00.950)       0:02:34.354 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Tuesday 13 January 2026  00:46:23 +0000 (0:00:00.968)       0:02:35.323 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Tuesday 13 January 2026  00:46:23 +0000 (0:00:00.147)       0:02:35.471 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Tuesday 13 January 2026  00:46:23 +0000 (0:00:00.606)       0:02:36.077 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Tuesday 13 January 2026  00:46:24 +0000 (0:00:00.786)       0:02:36.864 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Tuesday 13 January 2026  00:46:25 +0000 (0:00:00.590)       0:02:37.455 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Tuesday 13 January 2026  00:46:27 +0000 (0:00:02.165)       0:02:39.620 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Tuesday 13 January 2026  00:46:27 +0000 (0:00:00.472)       0:02:40.093 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Tuesday 13 January 2026  00:46:28 +0000 (0:00:00.979)       0:02:41.072 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Tuesday 13 January 2026  00:46:29 +0000 (0:00:00.754)       0:02:41.827 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Tuesday 13 January 2026  00:46:30 +0000 (0:00:00.597)       0:02:42.424 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Tuesday 13 January 2026  00:46:30 +0000 (0:00:00.645)       0:02:43.070 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Tuesday 13 January 2026  00:46:31 +0000 (0:00:00.645)       0:02:43.715 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Tuesday 13 January 2026  00:46:32 +0000 (0:00:00.679)       0:02:44.395 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Tuesday 13 January 2026  00:46:32 +0000 (0:00:00.652)       0:02:45.047 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Tuesday 13 January 2026  00:46:33 +0000 (0:00:00.710)       0:02:45.757 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Tuesday 13 January 2026  00:46:34 +0000 (0:00:00.558)       0:02:46.316 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Tuesday 13 January 2026  00:46:35 +0000 (0:00:01.010)       0:02:47.326 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Tuesday 13 January 2026  00:46:36 +0000 (0:00:01.087)       0:02:48.414 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Tuesday 13 January 2026  00:46:43 +0000 (0:00:07.433)       0:02:55.847 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Tuesday 13 January 2026  00:46:44 +0000 (0:00:00.840)       0:02:56.687 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Tuesday 13 January 2026  00:46:45 +0000 (0:00:00.844)       0:02:57.532 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Tuesday 13 January 2026  00:46:46 +0000 (0:00:01.323)       0:02:58.855 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Tuesday 13 January 2026  00:46:47 +0000 (0:00:00.569)       0:02:59.425 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Tuesday 13 January 2026  00:46:48 +0000 (0:00:00.745)       0:03:00.170 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Tuesday 13 January 2026  00:46:48 +0000 (0:00:00.603)       0:03:00.774 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Tuesday 13 January 2026  00:46:49 +0000 (0:00:00.736)       0:03:01.510 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Tuesday 13 January 2026  00:46:50 +0000 (0:00:00.626)       0:03:02.137 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Tuesday 13 January 2026  00:46:51 +0000 (0:00:01.146)       0:03:03.283 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Tuesday 13 January 2026  00:46:51 +0000 (0:00:00.512)       0:03:03.795 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Tuesday 13 January 2026  00:46:52 +0000 (0:00:00.793)       0:03:04.589 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Tuesday 13 January 2026  00:46:55 +0000 (0:00:02.915)       0:03:07.504 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Tuesday 13 January 2026  00:46:56 +0000 (0:00:01.085)       0:03:08.590 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Tuesday 13 January 2026  00:46:57 +0000 (0:00:00.843)       0:03:09.433 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Tuesday 13 January 2026  00:46:58 +0000 (0:00:01.148)       0:03:10.582 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Tuesday 13 January 2026  00:46:59 +0000 (0:00:00.653)       0:03:11.235 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Tuesday 13 January 2026  00:46:59 +0000 (0:00:00.811)       0:03:12.047 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Tuesday 13 January 2026  00:47:00 +0000 (0:00:00.499)       0:03:12.546 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 13 January 2026  00:47:01 +0000 (0:00:00.698)       0:03:13.245 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 13 January 2026  00:47:01 +0000 (0:00:00.632)       0:03:13.877 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 13 January 2026  00:47:02 +0000 (0:00:00.786)       0:03:14.663 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 13 January 2026  00:47:03 +0000 (0:00:00.527)       0:03:15.191 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 13 January 2026  00:47:03 +0000 (0:00:00.756)       0:03:15.947 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 13 January 2026  00:47:04 +0000 (0:00:00.359)       0:03:16.307 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 13 January 2026  00:47:04 +0000 (0:00:00.356)       0:03:16.663 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 13 January 2026  00:47:04 +0000 (0:00:00.337)       0:03:17.000 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 13 January 2026  00:47:05 +0000 (0:00:00.500)       0:03:17.501 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Tuesday 13 January 2026  00:47:06 +0000 (0:00:01.414)       0:03:18.916 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:47:08 +0000 (0:00:02.179)       0:03:21.095 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Tuesday 13 January 2026  00:47:09 +0000 (0:00:00.964)       0:03:22.060 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:47:10 +0000 (0:00:00.851)       0:03:22.911 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Tuesday 13 January 2026  00:47:11 +0000 (0:00:00.269)       0:03:23.180 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:47:12 +0000 (0:00:01.317)       0:03:24.498 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:47:13 +0000 (0:00:00.629)       0:03:25.127 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Tuesday 13 January 2026  00:47:13 +0000 (0:00:00.360)       0:03:25.488 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Tuesday 13 January 2026  00:47:14 +0000 (0:00:01.080)       0:03:26.568 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:47:14 +0000 (0:00:00.386)       0:03:26.955 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.310)       0:03:27.266 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.193)       0:03:27.459 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.252)       0:03:27.712 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.181)       0:03:27.893 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Tuesday 13 January 2026  00:47:15 +0000 (0:00:00.195)       0:03:28.089 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Tuesday 13 January 2026  00:47:16 +0000 (0:00:00.110)       0:03:28.199 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Tuesday 13 January 2026  00:47:16 +0000 (0:00:00.530)       0:03:28.729 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Tuesday 13 January 2026  00:47:16 +0000 (0:00:00.175)       0:03:28.905 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.378)       0:03:29.283 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.295)       0:03:29.579 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.174)       0:03:29.754 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Tuesday 13 January 2026  00:47:17 +0000 (0:00:00.201)       0:03:29.956 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:47:18 +0000 (0:00:00.854)       0:03:30.810 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Tuesday 13 January 2026  00:47:18 +0000 (0:00:00.260)       0:03:31.070 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:47:20 +0000 (0:00:01.208)       0:03:32.278 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:47:20 +0000 (0:00:00.698)       0:03:32.976 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Tuesday 13 January 2026  00:47:21 +0000 (0:00:00.413)       0:03:33.389 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:47:21 +0000 (0:00:00.690)       0:03:34.080 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Tuesday 13 January 2026  00:47:22 +0000 (0:00:00.631)       0:03:34.711 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:47:23 +0000 (0:00:01.327)       0:03:36.039 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:47:24 +0000 (0:00:00.712)       0:03:36.752 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Tuesday 13 January 2026  00:47:25 +0000 (0:00:00.355)       0:03:37.108 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Tuesday 13 January 2026  00:47:25 +0000 (0:00:00.911)       0:03:38.020 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:47:26 +0000 (0:00:00.857)       0:03:38.878 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Tuesday 13 January 2026  00:47:27 +0000 (0:00:00.545)       0:03:39.424 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:47:28 +0000 (0:00:01.170)       0:03:40.594 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:47:29 +0000 (0:00:00.643)       0:03:41.238 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mon] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:47:29 +0000 (0:00:00.644)       0:03:41.882 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:47:30 +0000 (0:00:00.954)       0:03:42.837 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:47:31 +0000 (0:00:00.512)       0:03:43.349 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:47:32 +0000 (0:00:01.104)       0:03:44.454 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:47:32 +0000 (0:00:00.324)       0:03:44.779 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:47:33 +0000 (0:00:00.325)       0:03:45.104 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:47:33 +0000 (0:00:00.330)       0:03:45.435 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:47:34 +0000 (0:00:01.105)       0:03:46.540 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:47:34 +0000 (0:00:00.313)       0:03:46.854 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:47:35 +0000 (0:00:00.333)       0:03:47.187 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:47:35 +0000 (0:00:00.649)       0:03:47.837 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:47:36 +0000 (0:00:01.059)       0:03:48.896 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:47:37 +0000 (0:00:00.318)       0:03:49.214 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:47:37 +0000 (0:00:00.305)       0:03:49.520 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:47:37 +0000 (0:00:00.324)       0:03:49.844 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:47:38 +0000 (0:00:00.365)       0:03:50.210 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:47:38 +0000 (0:00:00.630)       0:03:50.840 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:47:39 +0000 (0:00:00.452)       0:03:51.293 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:47:39 +0000 (0:00:00.338)       0:03:51.632 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:47:39 +0000 (0:00:00.334)       0:03:51.966 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:47:40 +0000 (0:00:00.718)       0:03:52.684 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 13 January 2026  00:47:41 +0000 (0:00:00.538)       0:03:53.223 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Tuesday 13 January 2026  00:47:41 +0000 (0:00:00.340)       0:03:53.563 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Tuesday 13 January 2026  00:47:42 +0000 (0:00:00.923)       0:03:54.486 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Tuesday 13 January 2026  00:47:42 +0000 (0:00:00.164)       0:03:54.651 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Tuesday 13 January 2026  00:47:43 +0000 (0:00:01.169)       0:03:55.821 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Tuesday 13 January 2026  00:47:44 +0000 (0:00:00.482)       0:03:56.304 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Tuesday 13 January 2026  00:47:44 +0000 (0:00:00.444)       0:03:56.748 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Tuesday 13 January 2026  00:47:45 +0000 (0:00:01.189)       0:03:57.937 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Tuesday 13 January 2026  00:47:46 +0000 (0:00:00.597)       0:03:58.535 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Tuesday 13 January 2026  00:47:46 +0000 (0:00:00.552)       0:03:59.088 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Tuesday 13 January 2026  00:47:47 +0000 (0:00:00.698)       0:03:59.786 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Tuesday 13 January 2026  00:47:49 +0000 (0:00:01.347)       0:04:01.134 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Tuesday 13 January 2026  00:47:49 +0000 (0:00:00.646)       0:04:01.780 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Tuesday 13 January 2026  00:47:52 +0000 (0:00:03.055)       0:04:04.836 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Tuesday 13 January 2026  00:47:54 +0000 (0:00:01.562)       0:04:06.398 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Tuesday 13 January 2026  00:47:54 +0000 (0:00:00.526)       0:04:06.925 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Tuesday 13 January 2026  00:47:56 +0000 (0:00:01.471)       0:04:08.396 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Tuesday 13 January 2026  00:48:00 +0000 (0:00:03.806)       0:04:12.202 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Tuesday 13 January 2026  00:48:01 +0000 (0:00:01.256)       0:04:13.459 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Tuesday 13 January 2026  00:48:01 +0000 (0:00:00.304)       0:04:13.763 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Tuesday 13 January 2026  00:48:02 +0000 (0:00:01.005)       0:04:14.769 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Tuesday 13 January 2026  00:48:03 +0000 (0:00:00.537)       0:04:15.306 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Tuesday 13 January 2026  00:48:03 +0000 (0:00:00.317)       0:04:15.623 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Tuesday 13 January 2026  00:48:04 +0000 (0:00:00.717)       0:04:16.341 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Tuesday 13 January 2026  00:48:05 +0000 (0:00:01.548)       0:04:17.890 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Tuesday 13 January 2026  00:48:06 +0000 (0:00:01.196)       0:04:19.086 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Tuesday 13 January 2026  00:48:08 +0000 (0:00:01.952)       0:04:21.038 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Tuesday 13 January 2026  00:48:11 +0000 (0:00:02.234)       0:04:23.273 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Tuesday 13 January 2026  00:48:11 +0000 (0:00:00.738)       0:04:24.012 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Tuesday 13 January 2026  00:48:33 +0000 (0:00:21.752)       0:04:45.764 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Tuesday 13 January 2026  00:48:44 +0000 (0:00:11.068)       0:04:56.832 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Tuesday 13 January 2026  00:48:45 +0000 (0:00:00.571)       0:04:57.404 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__f30337ccff23b6379a097436e4a93a5b8d5b6a4e'}]) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:49:00 +0000 (0:00:15.229)       0:05:12.634 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Tuesday 13 January 2026  00:49:00 +0000 (0:00:00.313)       0:05:12.947 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:49:01 +0000 (0:00:00.742)       0:05:13.690 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Tuesday 13 January 2026  00:49:01 +0000 (0:00:00.319)       0:05:14.009 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:49:02 +0000 (0:00:00.395)       0:05:14.405 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:49:03 +0000 (0:00:01.238)       0:05:15.644 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mgr] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:49:04 +0000 (0:00:00.546)       0:05:16.190 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:49:04 +0000 (0:00:00.480)       0:05:16.671 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:49:05 +0000 (0:00:00.752)       0:05:17.423 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:49:06 +0000 (0:00:00.733)       0:05:18.157 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:49:06 +0000 (0:00:00.305)       0:05:18.463 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:49:06 +0000 (0:00:00.527)       0:05:18.991 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:49:07 +0000 (0:00:00.318)       0:05:19.309 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:49:08 +0000 (0:00:00.788)       0:05:20.097 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:49:08 +0000 (0:00:00.329)       0:05:20.427 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:49:08 +0000 (0:00:00.545)       0:05:20.973 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:49:09 +0000 (0:00:00.716)       0:05:21.689 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:49:10 +0000 (0:00:00.708)       0:05:22.397 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:49:10 +0000 (0:00:00.225)       0:05:22.623 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:49:10 +0000 (0:00:00.424)       0:05:23.048 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:49:11 +0000 (0:00:00.254)       0:05:23.303 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:49:11 +0000 (0:00:00.261)       0:05:23.565 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:49:11 +0000 (0:00:00.261)       0:05:23.827 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:49:12 +0000 (0:00:00.310)       0:05:24.137 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:49:12 +0000 (0:00:00.672)       0:05:24.810 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:49:13 +0000 (0:00:00.321)       0:05:25.131 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:49:13 +0000 (0:00:00.307)       0:05:25.439 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 13 January 2026  00:49:14 +0000 (0:00:00.742)       0:05:26.181 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Tuesday 13 January 2026  00:49:14 +0000 (0:00:00.584)       0:05:26.765 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Tuesday 13 January 2026  00:49:15 +0000 (0:00:00.498)       0:05:27.263 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Tuesday 13 January 2026  00:49:15 +0000 (0:00:00.677)       0:05:27.941 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Tuesday 13 January 2026  00:49:16 +0000 (0:00:00.515)       0:05:28.456 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Tuesday 13 January 2026  00:49:27 +0000 (0:00:10.759)       0:05:39.216 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Tuesday 13 January 2026  00:49:27 +0000 (0:00:00.361)       0:05:39.578 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 13 January 2026  00:49:29 +0000 (0:00:02.190)       0:05:41.768 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Tuesday 13 January 2026  00:49:30 +0000 (0:00:01.278)       0:05:43.047 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Tuesday 13 January 2026  00:49:31 +0000 (0:00:01.039)       0:05:44.087 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Tuesday 13 January 2026  00:49:32 +0000 (0:00:00.322)       0:05:44.409 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Tuesday 13 January 2026  00:49:32 +0000 (0:00:00.329)       0:05:44.739 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Tuesday 13 January 2026  00:49:33 +0000 (0:00:00.815)       0:05:45.554 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Tuesday 13 January 2026  00:49:33 +0000 (0:00:00.322)       0:05:45.876 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Tuesday 13 January 2026  00:49:34 +0000 (0:00:00.320)       0:05:46.197 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Tuesday 13 January 2026  00:49:34 +0000 (0:00:00.724)       0:05:46.922 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Tuesday 13 January 2026  00:49:36 +0000 (0:00:01.235)       0:05:48.158 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Tuesday 13 January 2026  00:49:37 +0000 (0:00:01.214)       0:05:49.373 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Tuesday 13 January 2026  00:49:39 +0000 (0:00:01.815)       0:05:51.188 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Tuesday 13 January 2026  00:49:41 +0000 (0:00:02.155)       0:05:53.343 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Tuesday 13 January 2026  00:49:41 +0000 (0:00:00.410)       0:05:53.754 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Tuesday 13 January 2026  00:50:17 +0000 (0:00:35.440)       0:06:29.195 ******* ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Tuesday 13 January 2026  00:50:18 +0000 (0:00:01.200)       0:06:30.396 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Tuesday 13 January 2026  00:50:18 +0000 (0:00:00.312)       0:06:30.708 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Tuesday 13 January 2026  00:50:18 +0000 (0:00:00.141)       0:06:30.850 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Tuesday 13 January 2026  00:50:25 +0000 (0:00:06.291)       0:06:37.141 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:50:30 +0000 (0:00:05.428)       0:06:42.569 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Tuesday 13 January 2026  00:50:31 +0000 (0:00:00.808)       0:06:43.378 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:50:32 +0000 (0:00:00.842)       0:06:44.220 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Tuesday 13 January 2026  00:50:32 +0000 (0:00:00.348)       0:06:44.568 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:50:33 +0000 (0:00:01.090)       0:06:45.659 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:50:34 +0000 (0:00:00.600)       0:06:46.259 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-osd] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:50:34 +0000 (0:00:00.815)       0:06:47.075 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:50:35 +0000 (0:00:00.501)       0:06:47.576 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:50:36 +0000 (0:00:00.804)       0:06:48.380 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:50:36 +0000 (0:00:00.346)       0:06:48.726 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:50:37 +0000 (0:00:00.712)       0:06:49.439 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:50:38 +0000 (0:00:00.825)       0:06:50.265 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:50:39 +0000 (0:00:01.482)       0:06:51.747 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:50:39 +0000 (0:00:00.314)       0:06:52.062 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:50:40 +0000 (0:00:00.333)       0:06:52.395 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:50:40 +0000 (0:00:00.325)       0:06:52.721 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:50:41 +0000 (0:00:00.996)       0:06:53.717 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:50:42 +0000 (0:00:00.782)       0:06:54.500 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:50:42 +0000 (0:00:00.319)       0:06:54.820 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:50:43 +0000 (0:00:00.302)       0:06:55.122 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:50:43 +0000 (0:00:00.592)       0:06:55.715 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:50:43 +0000 (0:00:00.346)       0:06:56.061 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:50:44 +0000 (0:00:00.348)       0:06:56.410 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:50:44 +0000 (0:00:00.300)       0:06:56.710 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:50:45 +0000 (0:00:00.537)       0:06:57.248 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:50:45 +0000 (0:00:00.307)       0:06:57.555 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:50:45 +0000 (0:00:00.352)       0:06:57.908 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Tuesday 13 January 2026  00:50:46 +0000 (0:00:00.770)       0:06:58.678 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 13 January 2026  00:50:46 +0000 (0:00:00.315)       0:06:58.994 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Tuesday 13 January 2026  00:50:47 +0000 (0:00:00.607)       0:06:59.601 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Tuesday 13 January 2026  00:50:48 +0000 (0:00:00.520)       0:07:00.121 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Tuesday 13 January 2026  00:50:48 +0000 (0:00:00.608)       0:07:00.730 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Tuesday 13 January 2026  00:50:48 +0000 (0:00:00.305)       0:07:01.036 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Tuesday 13 January 2026  00:50:49 +0000 (0:00:00.657)       0:07:01.693 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Tuesday 13 January 2026  00:50:49 +0000 (0:00:00.354)       0:07:02.048 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Tuesday 13 January 2026  00:50:55 +0000 (0:00:05.525)       0:07:07.574 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Tuesday 13 January 2026  00:50:55 +0000 (0:00:00.316)       0:07:07.890 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Tuesday 13 January 2026  00:50:56 +0000 (0:00:00.524)       0:07:08.414 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Tuesday 13 January 2026  00:50:57 +0000 (0:00:01.266)       0:07:09.681 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 13 January 2026  00:50:59 +0000 (0:00:02.105)       0:07:11.787 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Tuesday 13 January 2026  00:51:01 +0000 (0:00:01.328)       0:07:13.116 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Tuesday 13 January 2026  00:51:03 +0000 (0:00:02.308)       0:07:15.425 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Use ceph-volume to create osds] *******************************",
                                "Tuesday 13 January 2026  00:51:04 +0000 (0:00:00.826)       0:07:16.251 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-b9be54a9-cd9c-568c-9220-61b18da052d9', 'data_vg': 'ceph-b9be54a9-cd9c-568c-9220-61b18da052d9'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-e91d200a-cf56-55df-b2f8-08f15361112f', 'data_vg': 'ceph-e91d200a-cf56-55df-b2f8-08f15361112f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5', 'data_vg': 'ceph-11aa5137-b5aa-5373-b4c1-0bd5a429c1a5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-03961d85-1922-5669-8251-0ccc6cca9fac', 'data_vg': 'ceph-03961d85-1922-5669-8251-0ccc6cca9fac'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-7ebda4f6-7b50-59b0-8273-b291dd7d1677', 'data_vg': 'ceph-7ebda4f6-7b50-59b0-8273-b291dd7d1677'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-2b3e8737-91e3-53c0-9b3a-5288a4111b63', 'data_vg': 'ceph-2b3e8737-91e3-53c0-9b3a-5288a4111b63'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Tuesday 13 January 2026  00:51:46 +0000 (0:00:42.301)       0:07:58.553 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Tuesday 13 January 2026  00:51:46 +0000 (0:00:00.323)       0:07:58.876 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Tuesday 13 January 2026  00:51:47 +0000 (0:00:00.752)       0:07:59.629 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Tuesday 13 January 2026  00:51:48 +0000 (0:00:00.721)       0:08:00.351 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Tuesday 13 January 2026  00:51:51 +0000 (0:00:03.093)       0:08:03.444 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Tuesday 13 January 2026  00:51:52 +0000 (0:00:00.746)       0:08:04.191 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Tuesday 13 January 2026  00:51:53 +0000 (0:00:01.114)       0:08:05.306 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Tuesday 13 January 2026  00:51:54 +0000 (0:00:01.169)       0:08:06.475 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Tuesday 13 January 2026  00:51:56 +0000 (0:00:01.669)       0:08:08.144 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Tuesday 13 January 2026  00:51:56 +0000 (0:00:00.597)       0:08:08.742 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Tuesday 13 January 2026  00:51:56 +0000 (0:00:00.317)       0:08:09.059 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=3)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Tuesday 13 January 2026  00:51:58 +0000 (0:00:01.078)       0:08:10.138 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Tuesday 13 January 2026  00:52:00 +0000 (0:00:02.017)       0:08:12.155 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Tuesday 13 January 2026  00:52:03 +0000 (0:00:03.878)       0:08:16.034 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Tuesday 13 January 2026  00:52:07 +0000 (0:00:03.097)       0:08:19.132 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Tuesday 13 January 2026  00:52:19 +0000 (0:00:12.335)       0:08:31.467 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:52:20 +0000 (0:00:01.044)       0:08:32.512 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Tuesday 13 January 2026  00:52:20 +0000 (0:00:00.347)       0:08:32.859 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Tuesday 13 January 2026  00:52:21 +0000 (0:00:00.518)       0:08:33.378 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:52:22 +0000 (0:00:00.944)       0:08:34.322 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Tuesday 13 January 2026  00:52:22 +0000 (0:00:00.323)       0:08:34.646 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Tuesday 13 January 2026  00:52:22 +0000 (0:00:00.243)       0:08:34.889 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:00.307)       0:08:35.197 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:00.245)       0:08:35.442 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:00.225)       0:08:35.668 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:00.116)       0:08:35.785 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Tuesday 13 January 2026  00:52:23 +0000 (0:00:00.199)       0:08:35.984 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Tuesday 13 January 2026  00:52:24 +0000 (0:00:00.856)       0:08:36.840 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:52:25 +0000 (0:00:00.412)       0:08:37.252 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Tuesday 13 January 2026  00:52:25 +0000 (0:00:00.356)       0:08:37.609 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Tuesday 13 January 2026  00:52:25 +0000 (0:00:00.214)       0:08:37.823 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-crash] ***************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:52:26 +0000 (0:00:00.966)       0:08:38.789 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:52:27 +0000 (0:00:01.276)       0:08:40.066 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:52:29 +0000 (0:00:01.033)       0:08:41.100 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:52:30 +0000 (0:00:01.257)       0:08:42.358 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:52:31 +0000 (0:00:00.833)       0:08:43.191 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:52:32 +0000 (0:00:01.063)       0:08:44.255 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:52:33 +0000 (0:00:00.868)       0:08:45.124 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:52:34 +0000 (0:00:01.464)       0:08:46.589 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:52:35 +0000 (0:00:00.589)       0:08:47.179 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:52:36 +0000 (0:00:00.993)       0:08:48.173 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:52:37 +0000 (0:00:01.163)       0:08:49.336 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:52:38 +0000 (0:00:01.219)       0:08:50.556 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:52:39 +0000 (0:00:00.584)       0:08:51.140 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:52:39 +0000 (0:00:00.700)       0:08:51.841 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:52:40 +0000 (0:00:00.519)       0:08:52.360 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:52:41 +0000 (0:00:00.735)       0:08:53.096 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:52:41 +0000 (0:00:00.537)       0:08:53.634 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:52:42 +0000 (0:00:00.727)       0:08:54.361 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:52:42 +0000 (0:00:00.509)       0:08:54.871 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:52:43 +0000 (0:00:00.824)       0:08:55.695 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:52:44 +0000 (0:00:00.476)       0:08:56.172 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Tuesday 13 January 2026  00:52:45 +0000 (0:00:01.063)       0:08:57.236 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:04.212)       0:09:01.448 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Tuesday 13 January 2026  00:52:51 +0000 (0:00:02.054)       0:09:03.503 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Tuesday 13 January 2026  00:52:53 +0000 (0:00:01.895)       0:09:05.398 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Tuesday 13 January 2026  00:52:54 +0000 (0:00:01.138)       0:09:06.537 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Tuesday 13 January 2026  00:52:55 +0000 (0:00:01.049)       0:09:07.587 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Tuesday 13 January 2026  00:52:57 +0000 (0:00:01.816)       0:09:09.403 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************",
                                "Tuesday 13 January 2026  00:53:00 +0000 (0:00:03.410)       0:09:12.814 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******",
                                "Tuesday 13 January 2026  00:53:02 +0000 (0:00:01.388)       0:09:14.202 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************",
                                "Tuesday 13 January 2026  00:53:02 +0000 (0:00:00.743)       0:09:14.946 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******",
                                "Tuesday 13 January 2026  00:53:05 +0000 (0:00:02.606)       0:09:17.552 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mds] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:53:06 +0000 (0:00:00.924)       0:09:18.477 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:53:06 +0000 (0:00:00.441)       0:09:18.919 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:53:07 +0000 (0:00:00.665)       0:09:19.585 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:53:07 +0000 (0:00:00.298)       0:09:19.884 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:53:08 +0000 (0:00:00.662)       0:09:20.546 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:53:09 +0000 (0:00:00.998)       0:09:21.545 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:53:10 +0000 (0:00:00.718)       0:09:22.263 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:53:10 +0000 (0:00:00.333)       0:09:22.596 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:53:10 +0000 (0:00:00.321)       0:09:22.918 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:53:11 +0000 (0:00:00.580)       0:09:23.499 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:53:12 +0000 (0:00:00.704)       0:09:24.204 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:53:12 +0000 (0:00:00.786)       0:09:24.990 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:53:13 +0000 (0:00:00.318)       0:09:25.309 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:53:13 +0000 (0:00:00.574)       0:09:25.883 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:53:14 +0000 (0:00:00.342)       0:09:26.225 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:53:14 +0000 (0:00:00.338)       0:09:26.564 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:53:14 +0000 (0:00:00.335)       0:09:26.900 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:53:15 +0000 (0:00:00.667)       0:09:27.568 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:53:15 +0000 (0:00:00.317)       0:09:27.885 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:53:16 +0000 (0:00:00.333)       0:09:28.218 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:53:16 +0000 (0:00:00.317)       0:09:28.536 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Tuesday 13 January 2026  00:53:17 +0000 (0:00:00.823)       0:09:29.360 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule details] *********************",
                                "Tuesday 13 January 2026  00:53:17 +0000 (0:00:00.386)       0:09:29.746 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule name] ************************",
                                "Tuesday 13 January 2026  00:53:19 +0000 (0:00:02.127)       0:09:31.874 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create filesystem pools] **************************************",
                                "Tuesday 13 January 2026  00:53:20 +0000 (0:00:00.223)       0:09:32.097 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create ceph filesystem] ***************************************",
                                "Tuesday 13 January 2026  00:53:28 +0000 (0:00:08.560)       0:09:40.657 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Tuesday 13 January 2026  00:53:32 +0000 (0:00:03.786)       0:09:44.444 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Tuesday 13 January 2026  00:53:32 +0000 (0:00:00.581)       0:09:45.026 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Tuesday 13 January 2026  00:53:33 +0000 (0:00:01.009)       0:09:46.036 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 13 January 2026  00:53:36 +0000 (0:00:02.379)       0:09:48.415 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Tuesday 13 January 2026  00:53:37 +0000 (0:00:01.588)       0:09:50.004 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Tuesday 13 January 2026  00:53:41 +0000 (0:00:03.348)       0:09:53.352 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Tuesday 13 January 2026  00:53:41 +0000 (0:00:00.306)       0:09:53.659 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Tuesday 13 January 2026  00:53:42 +0000 (0:00:00.814)       0:09:54.474 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Tuesday 13 January 2026  00:53:42 +0000 (0:00:00.530)       0:09:55.004 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Tuesday 13 January 2026  00:53:44 +0000 (0:00:01.189)       0:09:56.194 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Tuesday 13 January 2026  00:53:45 +0000 (0:00:01.385)       0:09:57.579 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Tuesday 13 January 2026  00:53:47 +0000 (0:00:01.822)       0:09:59.401 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Tuesday 13 January 2026  00:53:49 +0000 (0:00:01.998)       0:10:01.400 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:53:50 +0000 (0:00:01.684)       0:10:03.085 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Tuesday 13 January 2026  00:53:51 +0000 (0:00:00.711)       0:10:03.796 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:53:52 +0000 (0:00:00.861)       0:10:04.658 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Tuesday 13 January 2026  00:53:52 +0000 (0:00:00.342)       0:10:05.000 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:53:54 +0000 (0:00:01.186)       0:10:06.186 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:53:55 +0000 (0:00:00.984)       0:10:07.171 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 13 January 2026  00:53:55 +0000 (0:00:00.878)       0:10:08.049 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 13 January 2026  00:53:56 +0000 (0:00:00.500)       0:10:08.550 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 13 January 2026  00:53:57 +0000 (0:00:00.873)       0:10:09.425 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 13 January 2026  00:53:57 +0000 (0:00:00.310)       0:10:09.736 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 13 January 2026  00:53:58 +0000 (0:00:00.680)       0:10:10.417 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 13 January 2026  00:53:59 +0000 (0:00:01.043)       0:10:11.461 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 13 January 2026  00:54:00 +0000 (0:00:00.739)       0:10:12.200 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 13 January 2026  00:54:00 +0000 (0:00:00.312)       0:10:12.512 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 13 January 2026  00:54:00 +0000 (0:00:00.334)       0:10:12.847 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 13 January 2026  00:54:01 +0000 (0:00:00.583)       0:10:13.430 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 13 January 2026  00:54:02 +0000 (0:00:00.754)       0:10:14.185 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 13 January 2026  00:54:02 +0000 (0:00:00.883)       0:10:15.069 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 13 January 2026  00:54:03 +0000 (0:00:00.337)       0:10:15.406 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 13 January 2026  00:54:03 +0000 (0:00:00.676)       0:10:16.082 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 13 January 2026  00:54:04 +0000 (0:00:00.379)       0:10:16.462 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 13 January 2026  00:54:04 +0000 (0:00:00.357)       0:10:16.820 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 13 January 2026  00:54:05 +0000 (0:00:00.320)       0:10:17.140 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 13 January 2026  00:54:05 +0000 (0:00:00.631)       0:10:17.772 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 13 January 2026  00:54:05 +0000 (0:00:00.303)       0:10:18.075 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 13 January 2026  00:54:06 +0000 (0:00:00.320)       0:10:18.396 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 13 January 2026  00:54:06 +0000 (0:00:00.320)       0:10:18.716 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Tuesday 13 January 2026  00:54:07 +0000 (0:00:00.808)       0:10:19.524 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Tuesday 13 January 2026  00:54:07 +0000 (0:00:00.520)       0:10:20.045 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 13 January 2026  00:54:10 +0000 (0:00:02.284)       0:10:22.329 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Tuesday 13 January 2026  00:54:11 +0000 (0:00:01.522)       0:10:23.852 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Tuesday 13 January 2026  00:54:12 +0000 (0:00:00.334)       0:10:24.186 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Tuesday 13 January 2026  00:54:12 +0000 (0:00:00.545)       0:10:24.732 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Tuesday 13 January 2026  00:54:14 +0000 (0:00:01.429)       0:10:26.161 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Tuesday 13 January 2026  00:54:17 +0000 (0:00:03.924)       0:10:30.085 ******* ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 13 January 2026  00:54:20 +0000 (0:00:02.193)       0:10:32.279 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Tuesday 13 January 2026  00:54:21 +0000 (0:00:01.082)       0:10:33.361 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Tuesday 13 January 2026  00:54:21 +0000 (0:00:00.229)       0:10:33.591 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Tuesday 13 January 2026  00:54:22 +0000 (0:00:01.148)       0:10:34.739 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Tuesday 13 January 2026  00:54:23 +0000 (0:00:00.594)       0:10:35.334 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Tuesday 13 January 2026  00:54:53 +0000 (0:00:30.369)       0:11:05.703 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Tuesday 13 January 2026  00:54:53 +0000 (0:00:00.327)       0:11:06.030 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Tuesday 13 January 2026  00:54:54 +0000 (0:00:00.295)       0:11:06.326 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Tuesday 13 January 2026  00:54:55 +0000 (0:00:00.806)       0:11:07.132 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Tuesday 13 January 2026  00:54:55 +0000 (0:00:00.559)       0:11:07.692 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Tuesday 13 January 2026  00:54:56 +0000 (0:00:01.367)       0:11:09.060 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Tuesday 13 January 2026  00:54:58 +0000 (0:00:01.575)       0:11:10.636 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Tuesday 13 January 2026  00:55:00 +0000 (0:00:02.274)       0:11:12.910 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 13 January 2026  00:55:03 +0000 (0:00:02.951)       0:11:15.862 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Tuesday 13 January 2026  00:55:04 +0000 (0:00:00.524)       0:11:16.386 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Tuesday 13 January 2026  00:55:04 +0000 (0:00:00.514)       0:11:16.901 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Tuesday 13 January 2026  00:55:05 +0000 (0:00:00.607)       0:11:17.508 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Tuesday 13 January 2026  00:55:05 +0000 (0:00:00.333)       0:11:17.841 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Tuesday 13 January 2026  00:55:06 +0000 (0:00:00.591)       0:11:18.433 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=31  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=34  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:55:06 +0000 (0:00:00.257)       0:11:18.690 ******* ",
                                "=============================================================================== ",
                                "ceph-container-common : Pulling Ceph container image ------------------- 60.89s",
                                "ceph-osd : Use ceph-volume to create osds ------------------------------ 42.30s",
                                "ceph-mgr : Wait for all mgr to be up ----------------------------------- 35.44s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 30.37s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.75s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 15.23s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 12.34s",
                                "ceph-mon : Fetch ceph initial keys ------------------------------------- 11.07s",
                                "ceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.76s",
                                "ceph-mds : Create filesystem pools -------------------------------------- 8.56s",
                                "ceph-config : Create ceph initial directories --------------------------- 7.43s",
                                "ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.29s",
                                "ceph-osd : Apply operating system tuning -------------------------------- 5.53s",
                                "ceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.43s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 5.02s",
                                "ceph-crash : Create client.crash keyring -------------------------------- 4.21s",
                                "ceph-rgw : Create rgw keyrings ------------------------------------------ 3.92s",
                                "ceph-osd : Systemd start osd -------------------------------------------- 3.88s",
                                "ceph-mon : Generate initial monmap -------------------------------------- 3.81s",
                                "ceph-mds : Create ceph filesystem --------------------------------------- 3.79s",
                                "\u001b[32m2026-01-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3a479429-e057-43b6-a348-a20efcea0e17 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Set kolla_action_mariadb] ************************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Tuesday 13 January 2026  00:52:29 +0000 (0:00:00.092)       0:00:00.092 ******* ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check MariaDB service] ***************************************************",
                                "Tuesday 13 January 2026  00:52:29 +0000 (0:00:00.051)       0:00:00.143 ******* ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********",
                                "Tuesday 13 January 2026  00:52:32 +0000 (0:00:03.026)       0:00:03.170 ******* ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = kolla_action_ng] ******************************",
                                "Tuesday 13 January 2026  00:52:32 +0000 (0:00:00.082)       0:00:03.252 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:52:33 +0000 (0:00:00.166)       0:00:03.419 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:52:33 +0000 (0:00:00.292)       0:00:03.712 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Tuesday 13 January 2026  00:52:33 +0000 (0:00:00.661)       0:00:04.374 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:52:34 +0000 (0:00:00.388)       0:00:04.762 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Tuesday 13 January 2026  00:52:35 +0000 (0:00:00.641)       0:00:05.403 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Tuesday 13 January 2026  00:52:38 +0000 (0:00:03.419)       0:00:08.823 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Tuesday 13 January 2026  00:52:39 +0000 (0:00:00.687)       0:00:09.510 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Tuesday 13 January 2026  00:52:40 +0000 (0:00:01.233)       0:00:10.744 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Tuesday 13 January 2026  00:52:44 +0000 (0:00:04.102)       0:00:14.846 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Tuesday 13 January 2026  00:52:45 +0000 (0:00:01.025)       0:00:15.872 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:52:48 +0000 (0:00:03.472)       0:00:19.345 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  00:52:49 +0000 (0:00:00.474)       0:00:19.819 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:52:52 +0000 (0:00:03.263)       0:00:23.082 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  00:52:55 +0000 (0:00:02.926)       0:00:26.009 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Check mariadb containers] **************************************",
                                "Tuesday 13 January 2026  00:52:58 +0000 (0:00:03.340)       0:00:29.350 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Tuesday 13 January 2026  00:53:02 +0000 (0:00:03.952)       0:00:33.302 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Tuesday 13 January 2026  00:53:03 +0000 (0:00:00.794)       0:00:34.097 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Tuesday 13 January 2026  00:53:04 +0000 (0:00:00.443)       0:00:34.540 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Tuesday 13 January 2026  00:53:04 +0000 (0:00:00.284)       0:00:34.825 ******* ",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Tuesday 13 January 2026  00:53:15 +0000 (0:00:10.939)       0:00:45.765 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Tuesday 13 January 2026  00:53:15 +0000 (0:00:00.406)       0:00:46.171 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Tuesday 13 January 2026  00:53:16 +0000 (0:00:00.641)       0:00:46.812 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Tuesday 13 January 2026  00:53:16 +0000 (0:00:00.422)       0:00:47.235 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Tuesday 13 January 2026  00:53:17 +0000 (0:00:00.407)       0:00:47.642 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Tuesday 13 January 2026  00:53:17 +0000 (0:00:00.396)       0:00:48.038 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:53:18 +0000 (0:00:00.642)       0:00:48.680 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [mariadb : Running MariaDB bootstrap container] ***************************",
                                "Tuesday 13 January 2026  00:53:18 +0000 (0:00:00.410)       0:00:49.091 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap host name into facts] **************************",
                                "Tuesday 13 January 2026  00:53:29 +0000 (0:00:10.549)       0:00:59.640 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:53:29 +0000 (0:00:00.130)       0:00:59.771 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Starting first MariaDB container] *******************",
                                "Tuesday 13 January 2026  00:53:30 +0000 (0:00:00.989)       0:01:00.760 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******",
                                "Tuesday 13 January 2026  00:53:38 +0000 (0:00:07.915)       0:01:08.676 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******",
                                "Tuesday 13 January 2026  00:53:39 +0000 (0:00:01.577)       0:01:10.253 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***",
                                "Tuesday 13 January 2026  00:53:42 +0000 (0:00:02.502)       0:01:12.756 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Tuesday 13 January 2026  00:53:42 +0000 (0:00:00.117)       0:01:12.874 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Tuesday 13 January 2026  00:53:42 +0000 (0:00:00.351)       0:01:13.225 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 13 January 2026  00:53:43 +0000 (0:00:00.599)       0:01:13.825 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 13 January 2026  00:54:01 +0000 (0:00:18.426)       0:01:32.252 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 13 January 2026  00:54:17 +0000 (0:00:15.633)       0:01:47.885 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 13 January 2026  00:54:19 +0000 (0:00:02.227)       0:01:50.113 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 13 January 2026  00:54:37 +0000 (0:00:17.340)       0:02:07.453 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 13 January 2026  00:54:52 +0000 (0:00:15.602)       0:02:23.056 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 13 January 2026  00:54:55 +0000 (0:00:02.492)       0:02:25.548 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 13 January 2026  00:55:12 +0000 (0:00:17.259)       0:02:42.807 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 13 January 2026  00:55:13 +0000 (0:00:00.653)       0:02:43.461 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Tuesday 13 January 2026  00:55:15 +0000 (0:00:02.666)       0:02:46.128 ******* ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Tuesday 13 January 2026  00:55:16 +0000 (0:00:00.555)       0:02:46.684 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Tuesday 13 January 2026  00:55:18 +0000 (0:00:02.256)       0:02:48.940 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Tuesday 13 January 2026  00:55:20 +0000 (0:00:02.080)       0:02:51.021 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Tuesday 13 January 2026  00:55:23 +0000 (0:00:02.638)       0:02:53.660 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Tuesday 13 January 2026  00:55:25 +0000 (0:00:02.407)       0:02:56.067 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Tuesday 13 January 2026  00:55:29 +0000 (0:00:03.538)       0:02:59.605 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:55:29 +0000 (0:00:00.232)       0:02:59.837 ******* ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 35.77s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 31.24s",
                                "mariadb : Restart MariaDB container ------------------------------------ 17.26s",
                                "mariadb : Check MariaDB service port liveness -------------------------- 10.94s",
                                "mariadb : Running MariaDB bootstrap container -------------------------- 10.55s",
                                "mariadb : Starting first MariaDB container ------------------------------ 7.92s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 4.72s",
                                "mariadb : Copying over config.json files for services ------------------- 4.10s",
                                "mariadb : Check mariadb containers -------------------------------------- 3.95s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 3.54s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 3.47s",
                                "mariadb : Ensuring config directories exist ----------------------------- 3.42s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 3.34s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 3.26s",
                                "Check MariaDB service --------------------------------------------------- 3.03s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.93s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.67s",
                                "mariadb : Creating database backup user and setting permissions --------- 2.64s",
                                "mariadb : Wait for first MariaDB service to sync WSREP ------------------ 2.50s",
                                "mariadb : Granting permissions on Mariabackup database to backup user --- 2.41s",
                                "\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6d840277-c60b-4757-be1d-e554d76c0c33 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.315)       0:00:00.315 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.262)       0:00:00.578 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m",
                                "",
                                "PLAY [Apply role horizon] ******************************************************",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.376)       0:00:00.954 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Ensuring config directories exist] *****************************",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.467)       0:00:01.422 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : Set empty custom policy] ***************************************",
                                "Tuesday 13 January 2026  00:55:36 +0000 (0:00:01.125)       0:00:02.547 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.371)       0:00:02.918 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.613)       0:00:03.531 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.263)       0:00:03.795 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.117)       0:00:03.913 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.425)       0:00:04.339 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.256)       0:00:04.596 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.120)       0:00:04.717 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:00.272)       0:00:04.989 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:00.298)       0:00:05.287 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:00.300)       0:00:05.588 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:00.259)       0:00:05.847 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:40 +0000 (0:00:00.284)       0:00:06.132 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:40 +0000 (0:00:00.106)       0:00:06.238 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:40 +0000 (0:00:00.248)       0:00:06.487 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.405)       0:00:06.892 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.130)       0:00:07.022 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.252)       0:00:07.275 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.253)       0:00:07.528 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.107)       0:00:07.636 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:41 +0000 (0:00:00.234)       0:00:07.870 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:42 +0000 (0:00:00.442)       0:00:08.312 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:42 +0000 (0:00:00.107)       0:00:08.419 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:42 +0000 (0:00:00.240)       0:00:08.660 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.258)       0:00:08.919 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.109)       0:00:09.028 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.370)       0:00:09.399 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.277)       0:00:09.676 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.104)       0:00:09.781 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 13 January 2026  00:55:44 +0000 (0:00:00.239)       0:00:10.020 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:55:44 +0000 (0:00:00.286)       0:00:10.307 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 13 January 2026  00:55:44 +0000 (0:00:00.128)       0:00:10.435 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over config.json files for services] *******************",
                                "Tuesday 13 January 2026  00:55:45 +0000 (0:00:00.520)       0:00:10.956 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over horizon.conf] *************************************",
                                "Tuesday 13 January 2026  00:55:46 +0000 (0:00:01.692)       0:00:12.648 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over kolla-settings.py] ********************************",
                                "Tuesday 13 January 2026  00:55:48 +0000 (0:00:01.716)       0:00:14.365 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom-settings.py] *******************************",
                                "Tuesday 13 January 2026  00:55:50 +0000 (0:00:02.318)       0:00:16.683 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over existing policy file] *****************************",
                                "Tuesday 13 January 2026  00:55:52 +0000 (0:00:02.063)       0:00:18.747 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom themes] ************************************",
                                "Tuesday 13 January 2026  00:55:53 +0000 (0:00:00.298)       0:00:19.045 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:55:53 +0000 (0:00:00.310)       0:00:19.356 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  00:55:54 +0000 (0:00:00.844)       0:00:20.201 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:55:55 +0000 (0:00:01.462)       0:00:21.664 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  00:55:56 +0000 (0:00:00.629)       0:00:22.293 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Deploy horizon container] **************************************",
                                "Tuesday 13 January 2026  00:55:57 +0000 (0:00:00.950)       0:00:23.243 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:55:59 +0000 (0:00:01.652)       0:00:24.896 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:55:59 +0000 (0:00:00.302)       0:00:25.198 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database] *************************************",
                                "Tuesday 13 January 2026  00:55:59 +0000 (0:00:00.516)       0:00:25.715 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database user and setting permissions] ********",
                                "Tuesday 13 January 2026  00:56:02 +0000 (0:00:02.774)       0:00:28.489 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Running Horizon bootstrap container] ***************************",
                                "Tuesday 13 January 2026  00:56:05 +0000 (0:00:03.164)       0:00:31.654 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:16.846)       0:00:48.501 ******* ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:00.075)       0:00:48.577 ******* ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:00.081)       0:00:48.658 ******* ",
                                "",
                                "RUNNING HANDLER [horizon : Restart horizon container] **************************",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:00.069)       0:00:48.727 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:57:16 +0000 (0:00:53.330)       0:01:42.058 ******* ",
                                "=============================================================================== ",
                                "horizon : Restart horizon container ------------------------------------ 53.33s",
                                "horizon : Running Horizon bootstrap container -------------------------- 16.85s",
                                "horizon : Creating Horizon database user and setting permissions -------- 3.17s",
                                "horizon : Creating Horizon database ------------------------------------- 2.77s",
                                "horizon : Copying over kolla-settings.py -------------------------------- 2.32s",
                                "horizon : Copying over custom-settings.py ------------------------------- 2.06s",
                                "horizon : Copying over horizon.conf ------------------------------------- 1.72s",
                                "horizon : Copying over config.json files for services ------------------- 1.69s",
                                "horizon : Deploy horizon container -------------------------------------- 1.65s",
                                "service-cert-copy : horizon | Copying over extra CA certificates -------- 1.46s",
                                "horizon : Ensuring config directories exist ----------------------------- 1.13s",
                                "service-cert-copy : horizon | Copying over backend internal TLS key ----- 0.95s",
                                "horizon : include_tasks ------------------------------------------------- 0.84s",
                                "service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.63s",
                                "horizon : include_tasks ------------------------------------------------- 0.61s",
                                "horizon : Update custom policy file name -------------------------------- 0.52s",
                                "horizon : include_tasks ------------------------------------------------- 0.52s",
                                "horizon : include_tasks ------------------------------------------------- 0.47s",
                                "horizon : Update policy file name --------------------------------------- 0.44s",
                                "horizon : Update custom policy file name -------------------------------- 0.43s",
                                "\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e2829daa-db1e-4082-8441-642a23e938b6 is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Create ceph pools] *******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Tuesday 13 January 2026  00:55:11 +0000 (0:00:00.561)       0:00:00.561 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Tuesday 13 January 2026  00:55:12 +0000 (0:00:00.550)       0:00:01.111 ******* ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Tuesday 13 January 2026  00:55:12 +0000 (0:00:00.625)       0:00:01.737 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Tuesday 13 January 2026  00:55:12 +0000 (0:00:00.252)       0:00:01.989 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Tuesday 13 January 2026  00:55:13 +0000 (0:00:00.841)       0:00:02.831 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Tuesday 13 January 2026  00:55:14 +0000 (0:00:00.284)       0:00:03.116 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Tuesday 13 January 2026  00:55:14 +0000 (0:00:00.249)       0:00:03.365 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Tuesday 13 January 2026  00:55:14 +0000 (0:00:00.273)       0:00:03.639 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Tuesday 13 January 2026  00:55:15 +0000 (0:00:00.502)       0:00:04.142 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Tuesday 13 January 2026  00:55:15 +0000 (0:00:00.313)       0:00:04.456 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Tuesday 13 January 2026  00:55:16 +0000 (0:00:00.637)       0:00:05.093 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Tuesday 13 January 2026  00:55:16 +0000 (0:00:00.436)       0:00:05.529 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Tuesday 13 January 2026  00:55:18 +0000 (0:00:02.055)       0:00:07.585 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Tuesday 13 January 2026  00:55:19 +0000 (0:00:00.657)       0:00:08.242 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Tuesday 13 January 2026  00:55:20 +0000 (0:00:00.825)       0:00:09.068 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Tuesday 13 January 2026  00:55:20 +0000 (0:00:00.417)       0:00:09.486 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '000f2103ee99', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-01-13 00:55:17.152761', 'end': '2026-01-13 00:55:17.180036', 'delta': '0:00:00.027275', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['000f2103ee99'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '84ac1be37239', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-01-13 00:55:17.884463', 'end': '2026-01-13 00:55:17.909365', 'delta': '0:00:00.024902', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['84ac1be37239'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '01f2672e35a4', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-01-13 00:55:18.404904', 'end': '2026-01-13 00:55:18.432868', 'delta': '0:00:00.027964', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['01f2672e35a4'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Tuesday 13 January 2026  00:55:20 +0000 (0:00:00.198)       0:00:09.685 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Tuesday 13 January 2026  00:55:21 +0000 (0:00:00.439)       0:00:10.124 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Tuesday 13 January 2026  00:55:22 +0000 (0:00:01.873)       0:00:11.998 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Tuesday 13 January 2026  00:55:23 +0000 (0:00:00.295)       0:00:12.294 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 13 January 2026  00:55:23 +0000 (0:00:00.391)       0:00:12.685 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Tuesday 13 January 2026  00:55:24 +0000 (0:00:00.489)       0:00:13.174 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Tuesday 13 January 2026  00:55:24 +0000 (0:00:00.113)       0:00:13.288 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 13 January 2026  00:55:24 +0000 (0:00:00.233)       0:00:13.521 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Tuesday 13 January 2026  00:55:24 +0000 (0:00:00.285)       0:00:13.807 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Tuesday 13 January 2026  00:55:25 +0000 (0:00:00.310)       0:00:14.117 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Tuesday 13 January 2026  00:55:25 +0000 (0:00:00.528)       0:00:14.646 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Tuesday 13 January 2026  00:55:25 +0000 (0:00:00.311)       0:00:14.957 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Tuesday 13 January 2026  00:55:26 +0000 (0:00:00.307)       0:00:15.265 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Tuesday 13 January 2026  00:55:26 +0000 (0:00:00.322)       0:00:15.587 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Tuesday 13 January 2026  00:55:27 +0000 (0:00:00.564)       0:00:16.151 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Tuesday 13 January 2026  00:55:27 +0000 (0:00:00.720)       0:00:16.872 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9', 'dm-uuid-LVM-tI9LueIqoznnHWvc67dyxcKb2DRlZadxhD8MTBDVbVSuVr75iGA0ykjhJTLhvbvd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac', 'dm-uuid-LVM-GHCgDfhjqHbxrN6X57Au2JxG0UkZVV6SYAZhc8KzmZuq1WeEWDc3uD3fnm7izynW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5', 'dm-uuid-LVM-hgtH6tpzhnx2QQztd0bAxtrFNuWF2rUJ5NeecY0iboAd4WuXz2J4zhiyU5ciBGer'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63', 'dm-uuid-LVM-xy47BTMmezzKuhVgeOBsrflxsh2nMMZxq1yfesNVZn38knC8hXtIcPF2l4aSiZtk'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16', 'scsi-SQEMU_QEMU_HARDDISK_ffeaaf24-9754-44c8-bb36-eb3a5d2d5315-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b9be54a9--cd9c--568c--9220--61b18da052d9-osd--block--b9be54a9--cd9c--568c--9220--61b18da052d9'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-sZILOh-5tjd-Njbz-niJz-MLcH-ddwd-N90s5N', 'scsi-0QEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98', 'scsi-SQEMU_QEMU_HARDDISK_49cd33e4-72cd-4f3f-940d-55c9f0f00a98'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--03961d85--1922--5669--8251--0ccc6cca9fac-osd--block--03961d85--1922--5669--8251--0ccc6cca9fac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uzsK8R-6Gxn-oDpB-2Hms-tH0u-G7ac-aGEaLg', 'scsi-0QEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573', 'scsi-SQEMU_QEMU_HARDDISK_1f00cc32-4927-4d99-9c1e-b649b1d1f573'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e', 'scsi-SQEMU_QEMU_HARDDISK_0a292857-8cd9-4a14-95ba-a5d022f4a90e'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-10-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16', 'scsi-SQEMU_QEMU_HARDDISK_5f6d3b65-3844-4001-8889-d6deb3f0644d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f', 'dm-uuid-LVM-xweh1YC5RDiVWhdx1PKskF5JCr6mh2cIruH8cXC0TzdCdfDxhyAoa4ykUz6BhD3x'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5-osd--block--11aa5137--b5aa--5373--b4c1--0bd5a429c1a5'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-6DKvSU-Cdbw-CbUk-lrwG-gfma-BvTf-I6WE2Y', 'scsi-0QEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056', 'scsi-SQEMU_QEMU_HARDDISK_6ad71b9e-76db-4ac5-b372-050f59253056'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677', 'dm-uuid-LVM-qXJ0ZdEvWcXk2vDlKmzolqGpgokmwsYUBrLx3bsgLllWFSDGo0grKruMjv28g8BC'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--2b3e8737--91e3--53c0--9b3a--5288a4111b63-osd--block--2b3e8737--91e3--53c0--9b3a--5288a4111b63'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-U2UOZW-SjBW-buGp-55CV-6Kqk-QEzG-AXKRcv', 'scsi-0QEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78', 'scsi-SQEMU_QEMU_HARDDISK_9db8234e-f6a8-4211-a809-87a509109e78'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c', 'scsi-SQEMU_QEMU_HARDDISK_5c0bff01-3898-4d25-903e-2ecdf087243c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-11-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16', 'scsi-SQEMU_QEMU_HARDDISK_306cfbe9-242f-441d-bc49-37fa1b1f4569-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e91d200a--cf56--55df--b2f8--08f15361112f-osd--block--e91d200a--cf56--55df--b2f8--08f15361112f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-FH2r3c-Cf2J-ryeq-ItYe-hsKy-vARI-3t2Zip', 'scsi-0QEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748', 'scsi-SQEMU_QEMU_HARDDISK_79922d84-0445-4535-976b-32e74e35a748'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7ebda4f6--7b50--59b0--8273--b291dd7d1677-osd--block--7ebda4f6--7b50--59b0--8273--b291dd7d1677'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-xHcG0E-vZHx-JCHk-rp13-0i6I-R8mG-hkrVOO', 'scsi-0QEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336', 'scsi-SQEMU_QEMU_HARDDISK_f69e02e7-d854-4ded-bb8d-51d0e0400336'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae', 'scsi-SQEMU_QEMU_HARDDISK_5295d09e-fddd-4452-8a25-9ba23e2b95ae'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-01-13-00-03-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Tuesday 13 January 2026  00:55:28 +0000 (0:00:00.562)       0:00:17.434 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Tuesday 13 January 2026  00:55:29 +0000 (0:00:00.656)       0:00:18.090 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 13 January 2026  00:55:29 +0000 (0:00:00.541)       0:00:18.632 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 13 January 2026  00:55:30 +0000 (0:00:00.613)       0:00:19.246 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 13 January 2026  00:55:30 +0000 (0:00:00.310)       0:00:19.556 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 13 January 2026  00:55:30 +0000 (0:00:00.429)       0:00:19.985 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Tuesday 13 January 2026  00:55:31 +0000 (0:00:00.547)       0:00:20.533 ******* ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Tuesday 13 January 2026  00:55:32 +0000 (0:00:01.011)       0:00:21.545 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Tuesday 13 January 2026  00:55:32 +0000 (0:00:00.370)       0:00:21.915 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 13 January 2026  00:55:33 +0000 (0:00:00.799)       0:00:22.715 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 13 January 2026  00:55:33 +0000 (0:00:00.303)       0:00:23.019 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.379)       0:00:23.398 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.257)       0:00:23.655 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.495)       0:00:24.151 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.345)       0:00:24.497 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.311)       0:00:24.808 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 13 January 2026  00:55:36 +0000 (0:00:00.293)       0:00:25.101 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 13 January 2026  00:55:36 +0000 (0:00:00.248)       0:00:25.350 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Tuesday 13 January 2026  00:55:36 +0000 (0:00:00.451)       0:00:25.801 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.820)       0:00:26.622 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Include tasks from the ceph-osd role] ************************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:01.620)       0:00:28.242 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [create openstack pool(s)] ************************************************",
                                "Tuesday 13 January 2026  00:55:39 +0000 (0:00:00.376)       0:00:28.619 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [generate keys] ***********************************************************",
                                "Tuesday 13 January 2026  00:56:26 +0000 (0:00:46.576)       0:01:15.195 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [get keys from monitors] **************************************************",
                                "Tuesday 13 January 2026  00:56:49 +0000 (0:00:23.636)       0:01:38.832 ******* ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [copy ceph key(s) if needed] **********************************************",
                                "Tuesday 13 January 2026  00:57:01 +0000 (0:00:12.096)       0:01:50.929 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:57:19 +0000 (0:00:17.775)       0:02:08.704 ******* ",
                                "=============================================================================== ",
                                "create openstack pool(s) ----------------------------------------------- 46.58s",
                                "generate keys ---------------------------------------------------------- 23.64s",
                                "copy ceph key(s) if needed --------------------------------------------- 17.78s",
                                "get keys from monitors ------------------------------------------------- 12.10s",
                                "ceph-facts : Find a running mon container ------------------------------- 2.06s",
                                "ceph-facts : Get current fsid if cluster is already running ------------- 1.87s",
                                "ceph-facts : Set_fact ceph_admin_command -------------------------------- 1.62s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 1.01s",
                                "ceph-facts : Check if podman binary is present -------------------------- 0.84s",
                                "ceph-facts : Check if the ceph mon socket is in-use --------------------- 0.83s",
                                "ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.82s",
                                "ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.80s",
                                "ceph-facts : Collect existed devices ------------------------------------ 0.72s",
                                "ceph-facts : Check for a ceph mon socket -------------------------------- 0.66s",
                                "ceph-facts : Check if the ceph conf exists ------------------------------ 0.66s",
                                "ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.64s",
                                "ceph-facts : Check if it is atomic host --------------------------------- 0.63s",
                                "ceph-facts : Read osd pool default crush rule --------------------------- 0.61s",
                                "ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks --- 0.56s",
                                "ceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.56s",
                                "\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:57:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c115080c-bb95-4c64-9141-3881243c5ded is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d32e32cc-419a-40cc-bd27-890c92e82cbf is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy ceph keys to the configuration repository] **************************",
                                "",
                                "TASK [Check if ceph keys exist] ************************************************",
                                "Tuesday 13 January 2026  00:57:24 +0000 (0:00:00.114)       0:00:00.114 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Fetch all ceph keys] *****************************************************",
                                "Tuesday 13 January 2026  00:57:28 +0000 (0:00:04.439)       0:00:04.553 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Create share directory] **************************************************",
                                "Tuesday 13 January 2026  00:57:33 +0000 (0:00:04.412)       0:00:08.966 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the share directory] **********************************",
                                "Tuesday 13 January 2026  00:57:34 +0000 (0:00:01.028)       0:00:09.994 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Check if target directories exist] ***************************************",
                                "Tuesday 13 January 2026  00:57:47 +0000 (0:00:13.266)       0:00:23.261 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the configuration directory] **************************",
                                "Tuesday 13 January 2026  00:57:51 +0000 (0:00:04.075)       0:00:27.336 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:57:58 +0000 (0:00:07.208)       0:00:34.545 ******* ",
                                "=============================================================================== ",
                                "Write ceph keys to the share directory --------------------------------- 13.27s",
                                "Write ceph keys to the configuration directory -------------------------- 7.21s",
                                "Check if ceph keys exist ------------------------------------------------ 4.44s",
                                "Fetch all ceph keys ----------------------------------------------------- 4.41s",
                                "Check if target directories exist --------------------------------------- 4.08s",
                                "Create share directory -------------------------------------------------- 1.03s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.256)       0:00:00.256 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.232)       0:00:00.488 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Apply role keystone] *****************************************************",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:55:34 +0000 (0:00:00.355)       0:00:00.843 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Ensuring config directories exist] ****************************",
                                "Tuesday 13 January 2026  00:55:35 +0000 (0:00:00.457)       0:00:01.301 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Check if policies shall be overwritten] ***********************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:01.705)       0:00:03.006 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Set keystone policy file] *************************************",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.112)       0:00:03.118 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Check if Keystone domain-specific config is supplied] *********",
                                "Tuesday 13 January 2026  00:55:37 +0000 (0:00:00.390)       0:00:03.509 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.738)       0:00:04.248 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over extra CA certificates] *******",
                                "Tuesday 13 January 2026  00:55:38 +0000 (0:00:00.499)       0:00:04.747 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:55:42 +0000 (0:00:03.370)       0:00:08.118 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****",
                                "Tuesday 13 January 2026  00:55:42 +0000 (0:00:00.642)       0:00:08.760 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over config.json files for services] ******************",
                                "Tuesday 13 January 2026  00:55:43 +0000 (0:00:00.639)       0:00:09.400 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone.conf] ***********************************",
                                "Tuesday 13 January 2026  00:55:46 +0000 (0:00:03.447)       0:00:12.848 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying keystone-startup script for keystone] *****************",
                                "Tuesday 13 January 2026  00:55:52 +0000 (0:00:05.521)       0:00:18.370 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Create Keystone domain-specific config directory] *************",
                                "Tuesday 13 January 2026  00:55:53 +0000 (0:00:01.428)       0:00:19.798 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Get file list in custom domains folder] ***********************",
                                "Tuesday 13 January 2026  00:55:54 +0000 (0:00:00.599)       0:00:20.398 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying Keystone Domain specific settings] ********************",
                                "Tuesday 13 January 2026  00:55:54 +0000 (0:00:00.284)       0:00:20.682 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over existing policy file] ****************************",
                                "Tuesday 13 January 2026  00:55:55 +0000 (0:00:00.595)       0:00:21.278 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:55:55 +0000 (0:00:00.568)       0:00:21.847 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over wsgi-keystone.conf] ******************************",
                                "Tuesday 13 January 2026  00:55:56 +0000 (0:00:00.301)       0:00:22.148 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "",
                                "TASK [keystone : Checking whether keystone-paste.ini file exists] **************",
                                "Tuesday 13 January 2026  00:55:57 +0000 (0:00:01.800)       0:00:23.949 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone-paste.ini] ******************************",
                                "Tuesday 13 January 2026  00:55:58 +0000 (0:00:00.900)       0:00:24.849 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Generate the required cron jobs for the node] *****************",
                                "Tuesday 13 January 2026  00:55:59 +0000 (0:00:00.852)       0:00:25.701 ******* ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***",
                                "Tuesday 13 January 2026  00:56:00 +0000 (0:00:01.059)       0:00:26.761 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-fernet] ****************************",
                                "Tuesday 13 January 2026  00:56:01 +0000 (0:00:00.296)       0:00:27.058 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-ssh] *******************************",
                                "Tuesday 13 January 2026  00:56:09 +0000 (0:00:08.632)       0:00:35.690 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "",
                                "TASK [keystone : Check keystone containers] ************************************",
                                "Tuesday 13 January 2026  00:56:12 +0000 (0:00:03.106)       0:00:38.797 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:56:15 +0000 (0:00:02.376)       0:00:41.173 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Creating keystone database] ***********************************",
                                "Tuesday 13 January 2026  00:56:15 +0000 (0:00:00.294)       0:00:41.467 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating Keystone database user and setting permissions] ******",
                                "Tuesday 13 January 2026  00:56:17 +0000 (0:00:02.494)       0:00:43.961 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Checking for any running keystone_fernet containers] **********",
                                "Tuesday 13 January 2026  00:56:20 +0000 (0:00:02.557)       0:00:46.519 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [keystone : Group nodes where keystone_fernet is running] *****************",
                                "Tuesday 13 January 2026  00:56:21 +0000 (0:00:01.208)       0:00:47.727 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:00.305)       0:00:48.033 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone bootstrap container] *************************",
                                "Tuesday 13 January 2026  00:56:22 +0000 (0:00:00.347)       0:00:48.381 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone fernet bootstrap container] ******************",
                                "Tuesday 13 January 2026  00:56:36 +0000 (0:00:14.013)       0:01:02.394 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:56:47 +0000 (0:00:11.244)       0:01:13.638 ******* ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:56:47 +0000 (0:00:00.065)       0:01:13.704 ******* ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:56:47 +0000 (0:00:00.077)       0:01:13.781 ******* ",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-ssh container] ********************",
                                "Tuesday 13 January 2026  00:56:47 +0000 (0:00:00.066)       0:01:13.848 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-fernet container] *****************",
                                "Tuesday 13 January 2026  00:57:03 +0000 (0:00:15.296)       0:01:29.145 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone container] ************************",
                                "Tuesday 13 January 2026  00:57:08 +0000 (0:00:05.427)       0:01:34.572 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:57:19 +0000 (0:00:10.936)       0:01:45.509 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Waiting for Keystone SSH port to be UP] ***********************",
                                "Tuesday 13 January 2026  00:57:20 +0000 (0:00:00.798)       0:01:46.307 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Run key distribution] *****************************************",
                                "Tuesday 13 January 2026  00:57:21 +0000 (0:00:00.812)       0:01:47.120 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating admin project, user, role, service, and endpoint] ****",
                                "Tuesday 13 January 2026  00:57:22 +0000 (0:00:01.571)       0:01:48.691 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating services] **********************",
                                "Tuesday 13 January 2026  00:57:34 +0000 (0:00:12.222)       0:02:00.913 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating endpoints] *********************",
                                "Tuesday 13 January 2026  00:58:04 +0000 (0:00:29.303)       0:02:30.217 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating projects] **********************",
                                "Tuesday 13 January 2026  00:58:10 +0000 (0:00:06.295)       0:02:36.512 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating users] *************************",
                                "Tuesday 13 January 2026  00:58:10 +0000 (0:00:00.127)       0:02:36.640 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating roles] *************************",
                                "Tuesday 13 January 2026  00:58:10 +0000 (0:00:00.109)       0:02:36.749 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Granting user roles] ********************",
                                "Tuesday 13 January 2026  00:58:10 +0000 (0:00:00.136)       0:02:36.885 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating default user role] ***********************************",
                                "Tuesday 13 January 2026  00:58:11 +0000 (0:00:00.830)       0:02:37.716 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:58:15 +0000 (0:00:03.662)       0:02:41.378 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:58:15 +0000 (0:00:00.409)       0:02:41.788 ******* ",
                                "=============================================================================== ",
                                "service-ks-register : keystone | Creating services --------------------- 29.30s",
                                "keystone : Restart keystone-ssh container ------------------------------ 15.30s",
                                "keystone : Running Keystone bootstrap container ------------------------ 14.01s",
                                "keystone : Creating admin project, user, role, service, and endpoint --- 12.22s",
                                "keystone : Running Keystone fernet bootstrap container ----------------- 11.24s",
                                "keystone : Restart keystone container ---------------------------------- 10.94s",
                                "keystone : Copying files for keystone-fernet ---------------------------- 8.63s",
                                "service-ks-register : keystone | Creating endpoints --------------------- 6.30s",
                                "keystone : Copying over keystone.conf ----------------------------------- 5.52s",
                                "keystone : Restart keystone-fernet container ---------------------------- 5.43s",
                                "keystone : Creating default user role ----------------------------------- 3.66s",
                                "keystone : Copying over config.json files for services ------------------ 3.45s",
                                "service-cert-copy : keystone | Copying over extra CA certificates ------- 3.37s",
                                "keystone : Copying files for keystone-ssh ------------------------------- 3.11s",
                                "keystone : Creating Keystone database user and setting permissions ------ 2.56s",
                                "keystone : Creating keystone database ----------------------------------- 2.49s",
                                "keystone : Check keystone containers ------------------------------------ 2.38s",
                                "keystone : Copying over wsgi-keystone.conf ------------------------------ 1.80s",
                                "keystone : Ensuring config directories exist ---------------------------- 1.71s",
                                "keystone : Run key distribution ----------------------------------------- 1.57s",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 65f5c828-773f-4884-becd-75772195a52b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:58:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc2828b6-ec68-4a65-b245-786e5c13977b is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Tuesday 13 January 2026  00:58:03 +0000 (0:00:00.235)       0:00:00.235 ******* ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Tuesday 13 January 2026  00:58:03 +0000 (0:00:00.240)       0:00:00.475 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Tuesday 13 January 2026  00:58:04 +0000 (0:00:01.330)       0:00:01.806 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Tuesday 13 January 2026  00:58:06 +0000 (0:00:01.525)       0:00:03.332 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Tuesday 13 January 2026  00:58:07 +0000 (0:00:00.952)       0:00:04.284 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Tuesday 13 January 2026  00:58:08 +0000 (0:00:00.943)       0:00:05.228 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Tuesday 13 January 2026  00:58:43 +0000 (0:00:35.321)       0:00:40.549 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Tuesday 13 January 2026  00:58:47 +0000 (0:00:04.039)       0:00:44.589 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Tuesday 13 January 2026  00:58:48 +0000 (0:00:00.541)       0:00:45.130 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Tuesday 13 January 2026  00:58:48 +0000 (0:00:00.132)       0:00:45.263 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******",
                                "Tuesday 13 January 2026  00:58:48 +0000 (0:00:00.577)       0:00:45.841 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***",
                                "Tuesday 13 January 2026  00:58:51 +0000 (0:00:02.034)       0:00:47.876 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******",
                                "Tuesday 13 January 2026  00:58:51 +0000 (0:00:00.754)       0:00:48.630 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****",
                                "Tuesday 13 January 2026  00:58:52 +0000 (0:00:00.727)       0:00:49.357 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:58:54 +0000 (0:00:01.628)       0:00:50.986 ******* ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Manage cephclient service ------------------ 35.32s",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 4.04s",
                                "osism.services.cephclient : Restart cephclient service ------------------ 2.03s",
                                "osism.services.cephclient : Copy bash completion scripts ---------------- 1.63s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.53s",
                                "osism.services.cephclient : Create required directories ----------------- 1.33s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.95s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.94s",
                                "osism.services.cephclient : Ensure that all containers are up ----------- 0.75s",
                                "osism.services.cephclient : Wait for an healthy service ----------------- 0.73s",
                                "osism.services.cephclient : Include rook task --------------------------- 0.58s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.54s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.24s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.13s",
                                "",
                                "",
                                "PLAY [Download ironic ipa images] **********************************************",
                                "",
                                "TASK [Ensure the destination directory exists] *********************************",
                                "Tuesday 13 January 2026  00:58:21 +0000 (0:00:00.212)       0:00:00.212 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent initramfs] *****************************************",
                                "Tuesday 13 January 2026  00:58:22 +0000 (0:00:01.242)       0:00:01.455 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent kernel] ********************************************",
                                "Tuesday 13 January 2026  00:58:55 +0000 (0:00:32.735)       0:00:34.191 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (3 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:59:22 +0000 (0:00:27.198)       0:01:01.389 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:59:23 +0000 (0:00:00.699)       0:01:02.088 ******* ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m",
                                "",
                                "PLAY [Apply role ironic] *******************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  00:59:24 +0000 (0:00:00.985)       0:01:03.073 ******* ",
                                "=============================================================================== ",
                                "Download ironic-agent initramfs ---------------------------------------- 32.74s",
                                "Download ironic-agent kernel ------------------------------------------- 27.20s",
                                "Ensure the destination directory exists --------------------------------- 1.24s",
                                "Group hosts based on enabled services ----------------------------------- 0.99s",
                                "Group hosts based on Kolla action --------------------------------------- 0.70s",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b639ab41-f416-4eeb-9ecc-fa20fc59daf4 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:58:22 +0000 (0:00:00.346)       0:00:00.346 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.445)       0:00:00.791 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m",
                                "",
                                "PLAY [Apply role barbican] *****************************************************",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.536)       0:00:01.328 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating services] **********************",
                                "Tuesday 13 January 2026  00:58:24 +0000 (0:00:00.487)       0:00:01.816 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating endpoints] *********************",
                                "Tuesday 13 January 2026  00:58:28 +0000 (0:00:03.974)       0:00:05.791 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating projects] **********************",
                                "Tuesday 13 January 2026  00:58:35 +0000 (0:00:07.737)       0:00:13.528 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating users] *************************",
                                "Tuesday 13 January 2026  00:58:39 +0000 (0:00:03.390)       0:00:16.918 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating roles] *************************",
                                "Tuesday 13 January 2026  00:58:43 +0000 (0:00:04.393)       0:00:21.312 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Granting user roles] ********************",
                                "Tuesday 13 January 2026  00:59:01 +0000 (0:00:17.685)       0:00:38.997 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring config directories exist] ****************************",
                                "Tuesday 13 January 2026  00:59:06 +0000 (0:00:04.792)       0:00:43.789 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring vassals config directories exist] ********************",
                                "Tuesday 13 January 2026  00:59:08 +0000 (0:00:02.354)       0:00:46.144 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m",
                                "",
                                "TASK [barbican : Check if policies shall be overwritten] ***********************",
                                "Tuesday 13 January 2026  00:59:10 +0000 (0:00:01.742)       0:00:47.886 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Set barbican policy file] *************************************",
                                "Tuesday 13 January 2026  00:59:10 +0000 (0:00:00.102)       0:00:47.988 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:59:10 +0000 (0:00:00.376)       0:00:48.365 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over extra CA certificates] *******",
                                "Tuesday 13 January 2026  00:59:11 +0000 (0:00:00.452)       0:00:48.817 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:59:14 +0000 (0:00:03.018)       0:00:51.836 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****",
                                "Tuesday 13 January 2026  00:59:15 +0000 (0:00:01.274)       0:00:53.111 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over config.json files for services] ******************",
                                "Tuesday 13 January 2026  00:59:16 +0000 (0:00:01.434)       0:00:54.545 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api.ini] ********************************",
                                "Tuesday 13 January 2026  00:59:21 +0000 (0:00:04.191)       0:00:58.737 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [barbican : Checking whether barbican-api-paste.ini file exists] **********",
                                "Tuesday 13 January 2026  00:59:23 +0000 (0:00:02.499)       0:01:01.236 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api-paste.ini] **************************",
                                "Tuesday 13 January 2026  00:59:24 +0000 (0:00:01.097)       0:01:02.333 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican.conf] ***********************************",
                                "Tuesday 13 January 2026  00:59:25 +0000 (0:00:00.662)       0:01:02.996 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over existing policy file] ****************************",
                                "Tuesday 13 January 2026  00:59:35 +0000 (0:00:10.273)       0:01:13.269 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Check barbican containers] ************************************",
                                "Tuesday 13 January 2026  00:59:36 +0000 (0:00:01.211)       0:01:14.481 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  00:59:40 +0000 (0:00:03.821)       0:01:18.302 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database] ***********************************",
                                "Tuesday 13 January 2026  00:59:41 +0000 (0:00:00.474)       0:01:18.777 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database user and setting permissions] ******",
                                "Tuesday 13 January 2026  00:59:43 +0000 (0:00:02.494)       0:01:21.271 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Running barbican bootstrap container] *************************",
                                "Tuesday 13 January 2026  00:59:46 +0000 (0:00:02.649)       0:01:23.921 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:59:59 +0000 (0:00:13.110)       0:01:37.031 ******* ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:59:59 +0000 (0:00:00.121)       0:01:37.152 ******* ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 13 January 2026  00:59:59 +0000 (0:00:00.118)       0:01:37.271 ******* ",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-api container] ********************",
                                "Tuesday 13 January 2026  00:59:59 +0000 (0:00:00.066)       0:01:37.337 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******",
                                "Tuesday 13 January 2026  01:00:07 +0000 (0:00:07.277)       0:01:44.614 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-worker container] *****************",
                                "Tuesday 13 January 2026  01:00:12 +0000 (0:00:05.055)       0:01:49.671 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:00:20 +0000 (0:00:08.673)       0:01:58.345 ******* ",
                                "=============================================================================== ",
                                "service-ks-register : barbican | Creating roles ------------------------ 17.69s",
                                "barbican : Running barbican bootstrap container ------------------------ 13.11s",
                                "barbican : Copying over barbican.conf ---------------------------------- 10.27s",
                                "barbican : Restart barbican-worker container ---------------------------- 8.67s",
                                "service-ks-register : barbican | Creating endpoints --------------------- 7.74s",
                                "barbican : Restart barbican-api container ------------------------------- 7.28s",
                                "barbican : Restart barbican-keystone-listener container ----------------- 5.06s",
                                "service-ks-register : barbican | Granting user roles -------------------- 4.79s",
                                "service-ks-register : barbican | Creating users ------------------------- 4.39s",
                                "barbican : Copying over config.json files for services ------------------ 4.19s",
                                "service-ks-register : barbican | Creating services ---------------------- 3.97s",
                                "barbican : Check barbican containers ------------------------------------ 3.82s",
                                "service-ks-register : barbican | Creating projects ---------------------- 3.39s",
                                "service-cert-copy : barbican | Copying over extra CA certificates ------- 3.02s",
                                "barbican : Creating barbican database user and setting permissions ------ 2.65s",
                                "barbican : Copying over barbican-api.ini -------------------------------- 2.50s",
                                "barbican : Creating barbican database ----------------------------------- 2.49s",
                                "barbican : Ensuring config directories exist ---------------------------- 2.35s",
                                "barbican : Ensuring vassals config directories exist -------------------- 1.74s",
                                "service-cert-copy : barbican | Copying over backend internal TLS key ---- 1.43s",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5ffee94-dbe4-45cd-8561-7cdaa3c130e4 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 509960b6-ea70-4d2a-833f-c277c8679a2a is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Bootstraph ceph dashboard] ***********************************************",
                                "",
                                "TASK [Disable the ceph dashboard] **********************************************",
                                "Tuesday 13 January 2026  00:58:58 +0000 (0:00:00.287)       0:00:00.287 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/ssl to false] ******************************************",
                                "Tuesday 13 January 2026  00:59:00 +0000 (0:00:01.892)       0:00:02.179 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_port to 7000] ***********************************",
                                "Tuesday 13 January 2026  00:59:02 +0000 (0:00:01.350)       0:00:03.530 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************",
                                "Tuesday 13 January 2026  00:59:03 +0000 (0:00:01.031)       0:00:04.562 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_behaviour to error] ****************************",
                                "Tuesday 13 January 2026  00:59:04 +0000 (0:00:01.618)       0:00:06.180 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_error_status_code to 404] **********************",
                                "Tuesday 13 January 2026  00:59:06 +0000 (0:00:01.226)       0:00:07.407 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable the ceph dashboard] ***********************************************",
                                "Tuesday 13 January 2026  00:59:07 +0000 (0:00:01.012)       0:00:08.420 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Write ceph_dashboard_password to temporary file] *************************",
                                "Tuesday 13 January 2026  00:59:08 +0000 (0:00:01.121)       0:00:09.541 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create admin user] *******************************************************",
                                "Tuesday 13 January 2026  00:59:09 +0000 (0:00:01.194)       0:00:10.736 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Remove temporary file for ceph_dashboard_password] ***********************",
                                "Tuesday 13 January 2026  01:00:10 +0000 (0:01:00.865)       0:01:11.602 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 13 January 2026  01:00:10 +0000 (0:00:00.126)       0:01:11.728 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:11.353)       0:01:23.082 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 13 January 2026  01:00:32 +0000 (0:00:11.150)       0:01:34.232 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:00:34 +0000 (0:00:01.246)       0:01:35.479 ******* ",
                                "=============================================================================== ",
                                "Create admin user ------------------------------------------------------ 60.87s",
                                "Restart ceph manager service ------------------------------------------- 23.75s",
                                "Disable the ceph dashboard ---------------------------------------------- 1.89s",
                                "Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.62s",
                                "Set mgr/dashboard/ssl to false ------------------------------------------ 1.35s",
                                "Set mgr/dashboard/standby_behaviour to error ---------------------------- 1.23s",
                                "Write ceph_dashboard_password to temporary file ------------------------- 1.19s",
                                "Enable the ceph dashboard ----------------------------------------------- 1.12s",
                                "Set mgr/dashboard/server_port to 7000 ----------------------------------- 1.03s",
                                "Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.01s",
                                "Remove temporary file for ceph_dashboard_password ----------------------- 0.13s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:59:31 +0000 (0:00:00.215)       0:00:00.215 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:59:32 +0000 (0:00:00.227)       0:00:00.442 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m",
                                "",
                                "PLAY [Apply role placement] ****************************************************",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  00:59:32 +0000 (0:00:00.371)       0:00:00.814 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating services] *********************",
                                "Tuesday 13 January 2026  00:59:33 +0000 (0:00:00.693)       0:00:01.507 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating endpoints] ********************",
                                "Tuesday 13 January 2026  00:59:37 +0000 (0:00:04.225)       0:00:05.733 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating projects] *********************",
                                "Tuesday 13 January 2026  00:59:44 +0000 (0:00:06.768)       0:00:12.502 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating users] ************************",
                                "Tuesday 13 January 2026  00:59:47 +0000 (0:00:03.506)       0:00:16.009 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating roles] ************************",
                                "Tuesday 13 January 2026  00:59:52 +0000 (0:00:04.407)       0:00:20.416 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Granting user roles] *******************",
                                "Tuesday 13 January 2026  00:59:56 +0000 (0:00:04.351)       0:00:24.768 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:00:00 +0000 (0:00:04.456)       0:00:29.225 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Ensuring config directories exist] ***************************",
                                "Tuesday 13 January 2026  01:00:01 +0000 (0:00:00.528)       0:00:29.753 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Check if policies shall be overwritten] **********************",
                                "Tuesday 13 January 2026  01:00:02 +0000 (0:00:01.276)       0:00:31.030 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Set placement policy file] ***********************************",
                                "Tuesday 13 January 2026  01:00:02 +0000 (0:00:00.210)       0:00:31.240 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:00:03 +0000 (0:00:00.527)       0:00:31.768 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over extra CA certificates] ******",
                                "Tuesday 13 January 2026  01:00:03 +0000 (0:00:00.471)       0:00:32.240 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:00:05 +0000 (0:00:01.683)       0:00:33.923 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS key] ***",
                                "Tuesday 13 January 2026  01:00:06 +0000 (0:00:00.937)       0:00:34.861 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over config.json files for services] *****************",
                                "Tuesday 13 January 2026  01:00:07 +0000 (0:00:01.537)       0:00:36.398 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement.conf] *********************************",
                                "Tuesday 13 January 2026  01:00:09 +0000 (0:00:01.644)       0:00:38.043 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement-api wsgi configuration] ***************",
                                "Tuesday 13 January 2026  01:00:12 +0000 (0:00:03.354)       0:00:41.397 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [placement : Copying over migrate-db.rc.j2 configuration] *****************",
                                "Tuesday 13 January 2026  01:00:14 +0000 (0:00:01.679)       0:00:43.077 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over existing policy file] ***************************",
                                "Tuesday 13 January 2026  01:00:16 +0000 (0:00:01.726)       0:00:44.804 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Check placement containers] **********************************",
                                "Tuesday 13 January 2026  01:00:16 +0000 (0:00:00.434)       0:00:45.239 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases] ********************************",
                                "Tuesday 13 January 2026  01:00:17 +0000 (0:00:01.018)       0:00:46.257 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases user and setting permissions] ***",
                                "Tuesday 13 January 2026  01:00:20 +0000 (0:00:02.751)       0:00:49.009 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Running placement bootstrap container] ***********************",
                                "Tuesday 13 January 2026  01:00:22 +0000 (0:00:02.211)       0:00:51.221 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:36 +0000 (0:00:13.736)       0:01:04.957 ******* ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:36 +0000 (0:00:00.126)       0:01:05.084 ******* ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:36 +0000 (0:00:00.113)       0:01:05.198 ******* ",
                                "",
                                "RUNNING HANDLER [placement : Restart placement-api container] ******************",
                                "Tuesday 13 January 2026  01:00:36 +0000 (0:00:00.107)       0:01:05.306 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:00:44 +0000 (0:00:07.841)       0:01:13.147 ******* ",
                                "=============================================================================== ",
                                "placement : Running placement bootstrap container ---------------------- 13.74s",
                                "placement : Restart placement-api container ----------------------------- 7.84s",
                                "service-ks-register : placement | Creating endpoints -------------------- 6.77s",
                                "service-ks-register : placement | Granting user roles ------------------- 4.46s",
                                "service-ks-register : placement | Creating users ------------------------ 4.41s",
                                "service-ks-register : placement | Creating roles ------------------------ 4.35s",
                                "service-ks-register : placement | Creating services --------------------- 4.23s",
                                "service-ks-register : placement | Creating projects --------------------- 3.51s",
                                "placement : Copying over placement.conf --------------------------------- 3.36s",
                                "placement : Creating placement databases -------------------------------- 2.75s",
                                "placement : Creating placement databases user and setting permissions --- 2.21s",
                                "placement : Copying over migrate-db.rc.j2 configuration ----------------- 1.73s",
                                "service-cert-copy : placement | Copying over extra CA certificates ------ 1.68s",
                                "placement : Copying over placement-api wsgi configuration --------------- 1.68s",
                                "placement : Copying over config.json files for services ----------------- 1.64s",
                                "service-cert-copy : placement | Copying over backend internal TLS key --- 1.54s",
                                "placement : Ensuring config directories exist --------------------------- 1.28s",
                                "placement : Check placement containers ---------------------------------- 1.02s",
                                "service-cert-copy : placement | Copying over backend internal TLS certificate --- 0.94s",
                                "placement : include_tasks ----------------------------------------------- 0.69s",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dede57c8-30ae-496c-bc4c-7395ed5d10e8 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:00:49 +0000 (0:00:00.201)       0:00:00.201 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:00:49 +0000 (0:00:00.295)       0:00:00.496 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Keystone service] *******************************************",
                                "",
                                "TASK [Waiting for Keystone public port to be UP] *******************************",
                                "Tuesday 13 January 2026  01:00:50 +0000 (0:00:00.643)       0:00:01.139 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:00:51 +0000 (0:00:00.704)       0:00:01.844 ******* ",
                                "=============================================================================== ",
                                "Waiting for Keystone public port to be UP ------------------------------- 0.70s",
                                "Group hosts based on enabled services ----------------------------------- 0.64s",
                                "Group hosts based on Kolla action --------------------------------------- 0.30s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.296)       0:00:00.296 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.318)       0:00:00.615 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m",
                                "",
                                "PLAY [Apply role designate] ****************************************************",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.510)       0:00:01.125 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating services] *********************",
                                "Tuesday 13 January 2026  00:58:24 +0000 (0:00:00.507)       0:00:01.633 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating endpoints] ********************",
                                "Tuesday 13 January 2026  00:58:28 +0000 (0:00:04.383)       0:00:06.016 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating projects] *********************",
                                "Tuesday 13 January 2026  00:58:36 +0000 (0:00:07.449)       0:00:13.466 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating users] ************************",
                                "Tuesday 13 January 2026  00:58:39 +0000 (0:00:03.459)       0:00:16.925 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating roles] ************************",
                                "Tuesday 13 January 2026  00:58:44 +0000 (0:00:04.981)       0:00:21.907 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Granting user roles] *******************",
                                "Tuesday 13 January 2026  00:58:48 +0000 (0:00:03.586)       0:00:25.494 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m",
                                "",
                                "TASK [designate : Ensuring config directories exist] ***************************",
                                "Tuesday 13 January 2026  00:58:52 +0000 (0:00:04.196)       0:00:29.690 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Check if policies shall be overwritten] **********************",
                                "Tuesday 13 January 2026  00:58:55 +0000 (0:00:03.512)       0:00:33.203 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Set designate policy file] ***********************************",
                                "Tuesday 13 January 2026  00:58:56 +0000 (0:00:00.127)       0:00:33.330 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  00:58:56 +0000 (0:00:00.288)       0:00:33.619 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over extra CA certificates] ******",
                                "Tuesday 13 January 2026  00:58:57 +0000 (0:00:00.765)       0:00:34.385 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:59:03 +0000 (0:00:06.760)       0:00:41.146 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'reg\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9e2ad8df-a6a3-46e8-8666-5e5005fac8d4 is in state SUCCESS\u001b[0m",
                                "istry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS key] ***",
                                "Tuesday 13 January 2026  00:59:05 +0000 (0:00:01.991)       0:00:43.138 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over config.json files for services] *****************",
                                "Tuesday 13 January 2026  00:59:07 +0000 (0:00:02.090)       0:00:45.228 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over designate.conf] *********************************",
                                "Tuesday 13 January 2026  00:59:14 +0000 (0:00:06.970)       0:00:52.198 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over pools.yaml] *************************************",
                                "Tuesday 13 January 2026  00:59:37 +0000 (0:00:22.158)       0:01:14.357 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over named.conf] *************************************",
                                "Tuesday 13 January 2026  00:59:42 +0000 (0:00:05.526)       0:01:19.884 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.conf] **************************************",
                                "Tuesday 13 January 2026  00:59:45 +0000 (0:00:03.049)       0:01:22.933 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.key] ***************************************",
                                "Tuesday 13 January 2026  00:59:49 +0000 (0:00:03.988)       0:01:26.921 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  00:59:53 +0000 (0:00:03.645)       0:01:30.567 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over existing policy file] ***************************",
                                "Tuesday 13 January 2026  00:59:54 +0000 (0:00:00.783)       0:01:31.350 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Check designate containers] **********************************",
                                "Tuesday 13 January 2026  00:59:54 +0000 (0:00:00.716)       0:01:32.067 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:00:00 +0000 (0:00:05.229)       0:01:37.296 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases] ********************************",
                                "Tuesday 13 January 2026  01:00:00 +0000 (0:00:00.585)       0:01:37.881 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases user and setting permissions] ***",
                                "Tuesday 13 January 2026  01:00:03 +0000 (0:00:02.643)       0:01:40.525 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m",
                                "",
                                "TASK [designate : Running Designate bootstrap container] ***********************",
                                "Tuesday 13 January 2026  01:00:05 +0000 (0:00:02.537)       0:01:43.063 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:15.357)       0:01:58.421 ******* ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:00.063)       0:01:58.484 ******* ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:00.073)       0:01:58.558 ******* ",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:00.069)       0:01:58.627 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-api container] ******************",
                                "Tuesday 13 January 2026  01:00:34 +0000 (0:00:13.089)       0:02:11.717 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-central container] **************",
                                "Tuesday 13 January 2026  01:00:41 +0000 (0:00:06.992)       0:02:18.709 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-producer container] *************",
                                "Tuesday 13 January 2026  01:00:51 +0000 (0:00:09.724)       0:02:28.434 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-mdns container] *****************",
                                "Tuesday 13 January 2026  01:00:56 +0000 (0:00:05.027)       0:02:33.461 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-worker container] ***************",
                                "Tuesday 13 January 2026  01:01:01 +0000 (0:00:04.871)       0:02:38.333 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [designate : Non-destructive DNS pools update] ****************************",
                                "Tuesday 13 January 2026  01:01:07 +0000 (0:00:06.020)       0:02:44.353 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:01:14 +0000 (0:00:07.211)       0:02:51.565 ******* ",
                                "=============================================================================== ",
                                "designate : Copying over designate.conf -------------------------------- 22.16s",
                                "designate : Running Designate bootstrap container ---------------------- 15.36s",
                                "designate : Restart designate-backend-bind9 container ------------------ 13.09s",
                                "designate : Restart designate-central container ------------------------- 9.72s",
                                "service-ks-register : designate | Creating endpoints -------------------- 7.45s",
                                "designate : Non-destructive DNS pools update ---------------------------- 7.21s",
                                "designate : Restart designate-api container ----------------------------- 6.99s",
                                "designate : Copying over config.json files for services ----------------- 6.97s",
                                "service-cert-copy : designate | Copying over extra CA certificates ------ 6.76s",
                                "designate : Restart designate-worker container -------------------------- 6.02s",
                                "designate : Copying over pools.yaml ------------------------------------- 5.53s",
                                "designate : Check designate containers ---------------------------------- 5.23s",
                                "designate : Restart designate-producer container ------------------------ 5.03s",
                                "service-ks-register : designate | Creating users ------------------------ 4.98s",
                                "designate : Restart designate-mdns container ---------------------------- 4.87s",
                                "service-ks-register : designate | Creating services --------------------- 4.38s",
                                "service-ks-register : designate | Granting user roles ------------------- 4.20s",
                                "designate : Copying over rndc.conf -------------------------------------- 3.99s",
                                "designate : Copying over rndc.key --------------------------------------- 3.65s",
                                "service-ks-register : designate | Creating roles ------------------------ 3.59s",
                                "\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7ee72a7a-c81e-4cba-bad5-963c52381c77 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2d57ce62-ba5d-4095-a213-2e324d9777a2 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:01:20 +0000 (0:00:00.464)       0:00:00.464 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:01:22 +0000 (0:00:01.687)       0:00:02.151 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-rgw : include_tasks] ************************************************",
                                "Tuesday 13 January 2026  01:01:24 +0000 (0:00:02.389)       0:00:04.540 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating services] **********************",
                                "Tuesday 13 January 2026  01:01:29 +0000 (0:00:05.171)       0:00:09.712 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating endpoints] *********************",
                                "Tuesday 13 January 2026  01:01:33 +0000 (0:00:03.719)       0:00:13.433 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating projects] **********************",
                                "Tuesday 13 January 2026  01:01:40 +0000 (0:00:06.831)       0:00:20.265 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating users] *************************",
                                "Tuesday 13 January 2026  01:01:42 +0000 (0:00:02.703)       0:00:22.969 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating roles] *************************",
                                "Tuesday 13 January 2026  01:01:46 +0000 (0:00:03.351)       0:00:26.320 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Granting user roles] ********************",
                                "Tuesday 13 January 2026  01:01:52 +0000 (0:00:05.996)       0:00:32.316 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:01:57 +0000 (0:00:05.406)       0:00:37.723 ******* ",
                                "=============================================================================== ",
                                "service-ks-register : ceph-rgw | Creating endpoints --------------------- 6.83s",
                                "service-ks-register : ceph-rgw | Creating roles ------------------------- 6.00s",
                                "service-ks-register : ceph-rgw | Granting user roles -------------------- 5.41s",
                                "ceph-rgw : include_tasks ------------------------------------------------ 5.17s",
                                "service-ks-register : ceph-rgw | Creating services ---------------------- 3.72s",
                                "service-ks-register : ceph-rgw | Creating users ------------------------- 3.35s",
                                "service-ks-register : ceph-rgw | Creating projects ---------------------- 2.70s",
                                "Group hosts based on enabled services ----------------------------------- 2.39s",
                                "Group hosts based on Kolla action --------------------------------------- 1.69s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:00:27 +0000 (0:00:00.542)       0:00:00.542 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:00:28 +0000 (0:00:00.329)       0:00:00.871 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m",
                                "",
                                "PLAY [Apply role magnum] *******************************************************",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:00:28 +0000 (0:00:00.381)       0:00:01.253 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating services] ************************",
                                "Tuesday 13 January 2026  01:00:29 +0000 (0:00:00.604)       0:00:01.858 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating endpoints] ***********************",
                                "Tuesday 13 January 2026  01:00:33 +0000 (0:00:03.891)       0:00:05.749 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating projects] ************************",
                                "Tuesday 13 January 2026  01:00:39 +0000 (0:00:06.759)       0:00:12.508 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating users] ***************************",
                                "Tuesday 13 January 2026  01:00:42 +0000 (0:00:03.073)       0:00:15.581 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating roles] ***************************",
                                "Tuesday 13 January 2026  01:00:46 +0000 (0:00:03.983)       0:00:19.565 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Granting user roles] **********************",
                                "Tuesday 13 January 2026  01:00:50 +0000 (0:00:03.232)       0:00:22.797 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee domain] *********************************",
                                "Tuesday 13 January 2026  01:00:53 +0000 (0:00:03.616)       0:00:26.414 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user] ***********************************",
                                "Tuesday 13 January 2026  01:00:56 +0000 (0:00:03.250)       0:00:29.665 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user role] ******************************",
                                "Tuesday 13 January 2026  01:01:00 +0000 (0:00:03.956)       0:00:33.621 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  01:01:04 +0000 (0:00:03.186)       0:00:36.807 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Check if policies shall be overwritten] *************************",
                                "Tuesday 13 January 2026  01:01:05 +0000 (0:00:01.865)       0:00:38.673 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum policy file] *****************************************",
                                "Tuesday 13 January 2026  01:01:06 +0000 (0:00:00.125)       0:00:38.798 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check if kubeconfig file is supplied] ***************************",
                                "Tuesday 13 January 2026  01:01:06 +0000 (0:00:00.507)       0:00:39.305 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over kubeconfig file] ***********************************",
                                "Tuesday 13 January 2026  01:01:07 +0000 (0:00:00.884)       0:00:40.191 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum kubeconfig file's path] ******************************",
                                "Tuesday 13 January 2026  01:01:09 +0000 (0:00:02.464)       0:00:42.655 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:01:10 +0000 (0:00:00.289)       0:00:42.945 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over extra CA certificates] *********",
                                "Tuesday 13 January 2026  01:01:11 +0000 (0:00:00.892)       0:00:43.837 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:01:13 +0000 (0:00:02.435)       0:00:46.273 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******",
                                "Tuesday 13 January 2026  01:01:14 +0000 (0:00:01.132)       0:00:47.406 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  01:01:16 +0000 (0:00:01.693)       0:00:49.099 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over magnum.conf] ***************************************",
                                "Tuesday 13 January 2026  01:01:19 +0000 (0:00:02.940)       0:00:52.040 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over existing policy file] ******************************",
                                "Tuesday 13 January 2026  01:01:28 +0000 (0:00:09.650)       0:01:01.691 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check magnum containers] ****************************************",
                                "Tuesday 13 January 2026  01:01:30 +0000 (0:00:01.406)       0:01:03.098 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:01:35 +0000 (0:00:04.684)       0:01:07.782 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database] ***************************************",
                                "Tuesday 13 January 2026  01:01:36 +0000 (0:00:00.931)       0:01:08.714 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database user and setting permissions] **********",
                                "Tuesday 13 January 2026  01:01:38 +0000 (0:00:02.274)       0:01:10.989 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Running Magnum bootstrap container] *****************************",
                                "Tuesday 13 January 2026  01:01:40 +0000 (0:00:02.210)       0:01:13.199 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:01:54 +0000 (0:00:13.954)       0:01:27.154 ******* ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:01:54 +0000 (0:00:00.078)       0:01:27.232 ******* ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:01:54 +0000 (0:00:00.069)       0:01:27.302 ******* ",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-api container] ************************",
                                "Tuesday 13 January 2026  01:01:54 +0000 (0:00:00.072)       0:01:27.374 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-conductor container] ******************",
                                "Tuesday 13 January 2026  01:02:09 +0000 (0:00:14.827)       0:01:42.201 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:02:19 +0000 (0:00:09.627)       0:01:51.829 ******* ",
                                "=============================================================================== ",
                                "magnum : Restart magnum-api container ---------------------------------- 14.83s",
                                "magnum : Running Magnum bootstrap container ---------------------------- 13.95s",
                                "magnum : Copying over magnum.conf --------------------------------------- 9.65s",
                                "magnum : Restart magnum-conductor container ----------------------------- 9.63s",
                                "service-ks-register : magnum | Creating endpoints ----------------------- 6.76s",
                                "magnum : Check magnum containers ---------------------------------------- 4.68s",
                                "service-ks-register : magnum | Creating users --------------------------- 3.98s",
                                "magnum : Creating Magnum trustee user ----------------------------------- 3.96s",
                                "service-ks-register : magnum | Creating services ------------------------ 3.89s",
                                "service-ks-register : magnum | Granting user roles ---------------------- 3.62s",
                                "magnum : Creating Magnum trustee domain --------------------------------- 3.25s",
                                "service-ks-register : magnum | Creating roles --------------------------- 3.23s",
                                "magnum : Creating Magnum trustee user role ------------------------------ 3.19s",
                                "service-ks-register : magnum | Creating projects ------------------------ 3.07s",
                                "magnum : Copying over config.json files for services -------------------- 2.94s",
                                "magnum : Copying over kubeconfig file ----------------------------------- 2.46s",
                                "service-cert-copy : magnum | Copying over extra CA certificates --------- 2.44s",
                                "magnum : Creating Magnum database --------------------------------------- 2.27s",
                                "magnum : Creating Magnum database user and setting permissions ---------- 2.21s",
                                "magnum : Ensuring config directories exist ------------------------------ 1.87s",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  00:58:22 +0000 (0:00:00.277)       0:00:00.277 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:01.002)       0:00:01.280 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m",
                                "",
                                "PLAY [Apply role neutron] ******************************************************",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:58:23 +0000 (0:00:00.777)       0:00:02.058 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [neutron : Get container facts] *******************************************",
                                "Tuesday 13 January 2026  00:58:25 +0000 (0:00:01.060)       0:00:03.118 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Get container volume facts] ************************************",
                                "Tuesday 13 January 2026  00:58:26 +0000 (0:00:01.212)       0:00:04.331 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVN presence] ************************************",
                                "Tuesday 13 January 2026  00:58:27 +0000 (0:00:01.132)       0:00:05.463 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVS presence] ************************************",
                                "Tuesday 13 January 2026  00:58:28 +0000 (0:00:00.807)       0:00:06.270 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating services] ***********************",
                                "Tuesday 13 January 2026  00:58:28 +0000 (0:00:00.554)       0:00:06.825 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating endpoints] **********************",
                                "Tuesday 13 January 2026  00:58:32 +0000 (0:00:03.627)       0:00:10.453 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating projects] ***********************",
                                "Tuesday 13 January 2026  00:58:39 +0000 (0:00:07.329)       0:00:17.782 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating users] **************************",
                                "Tuesday 13 January 2026  00:58:43 +0000 (0:00:03.512)       0:00:21.295 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating roles] **************************",
                                "Tuesday 13 January 2026  00:58:47 +0000 (0:00:04.422)       0:00:25.717 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Granting user roles] *********************",
                                "Tuesday 13 January 2026  00:58:51 +0000 (0:00:03.684)       0:00:29.401 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:58:59 +0000 (0:00:07.706)       0:00:37.108 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Load and persist kernel modules] *****************************************",
                                "Tuesday 13 January 2026  00:58:59 +0000 (0:00:00.821)       0:00:37.930 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check IPv6 support] ********************************************",
                                "Tuesday 13 January 2026  00:59:02 +0000 (0:00:02.417)       0:00:40.347 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Tuesday 13 January 2026  00:59:04 +0000 (0:00:02.052)       0:00:42.399 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Ensuring config directories exist] *****************************",
                                "Tuesday 13 January 2026  00:59:07 +0000 (0:00:03.364)       0:00:45.764 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Check if extra ml2 plugins exists] *****************************",
                                "Tuesday 13 January 2026  00:59:11 +0000 (0:00:03.692)       0:00:49.456 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m",
                                "\u001b[1;35ma directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  00:59:12 +0000 (0:00:00.753)       0:00:50.210 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  00:59:13 +0000 (0:00:00.893)       0:00:51.103 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  00:59:15 +0000 (0:00:02.780)       0:00:53.884 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  00:59:18 +0000 (0:00:03.172)       0:00:57.056 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating TLS backend PEM File] *********************************",
                                "Tuesday 13 January 2026  00:59:21 +0000 (0:00:02.473)       0:00:59.529 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  00:59:24 +0000 (0:00:02.654)       0:01:02.184 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Set neutron policy file] ***************************************",
                                "Tuesday 13 January 2026  00:59:24 +0000 (0:00:00.181)       0:01:02.365 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over existing policy file] *****************************",
                                "Tuesday 13 January 2026  00:59:25 +0000 (0:00:00.733)       0:01:03.099 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_po\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a1dbb496-82c1-46c3-a715-dffbee1169f4 is in state SUCCESS\u001b[0m",
                                "rt neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over config.json files for services] *******************",
                                "Tuesday 13 January 2026  00:59:28 +0000 (0:00:03.295)       0:01:06.395 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron.conf] *************************************",
                                "Tuesday 13 January 2026  00:59:32 +0000 (0:00:04.234)       0:01:10.629 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_vpnaas.conf] ******************************",
                                "Tuesday 13 January 2026  00:59:39 +0000 (0:00:06.942)       0:01:17.571 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ssh key] ******************************************",
                                "Tuesday 13 January 2026  00:59:42 +0000 (0:00:02.536)       0:01:20.108 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ml2_conf.ini] *************************************",
                                "Tuesday 13 January 2026  00:59:44 +0000 (0:00:02.896)       0:01:23.005 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over linuxbridge_agent.ini] ****************************",
                                "Tuesday 13 January 2026  00:59:49 +0000 (0:00:04.323)       0:01:27.328 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over openvswitch_agent.ini] ****************************",
                                "Tuesday 13 January 2026  00:59:52 +0000 (0:00:02.794)       0:01:30.122 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over sriov_agent.ini] **********************************",
                                "Tuesday 13 January 2026  00:59:54 +0000 (0:00:02.490)       0:01:32.613 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over mlnx_agent.ini] ***********************************",
                                "Tuesday 13 January 2026  00:59:57 +0000 (0:00:02.603)       0:01:35.216 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over eswitchd.conf] ************************************",
                                "Tuesday 13 January 2026  00:59:58 +0000 (0:00:01.776)       0:01:36.993 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dhcp_agent.ini] ***********************************",
                                "Tuesday 13 January 2026  01:00:01 +0000 (0:00:02.683)       0:01:39.677 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dnsmasq.conf] *************************************",
                                "Tuesday 13 January 2026  01:00:03 +0000 (0:00:02.233)       0:01:41.910 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over l3_agent.ini] *************************************",
                                "Tuesday 13 January 2026  01:00:05 +0000 (0:00:02.159)       0:01:44.070 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over fwaas_driver.ini] *********************************",
                                "Tuesday 13 January 2026  01:00:08 +0000 (0:00:02.912)       0:01:46.983 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', \u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "'/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metadata_agent.ini] *******************************",
                                "Tuesday 13 January 2026  01:00:11 +0000 (0:00:02.437)       0:01:49.421 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************",
                                "Tuesday 13 January 2026  01:00:13 +0000 (0:00:02.472)       0:01:51.893 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metering_agent.ini] *******************************",
                                "Tuesday 13 January 2026  01:00:17 +0000 (0:00:03.433)       0:01:55.326 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ironic_neutron_agent.ini] *************************",
                                "Tuesday 13 January 2026  01:00:19 +0000 (0:00:01.857)       0:01:57.183 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over bgp_dragent.ini] **********************************",
                                "Tuesday 13 January 2026  01:00:21 +0000 (0:00:02.222)       0:01:59.406 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ovn_agent.ini] ************************************",
                                "Tuesday 13 January 2026  01:00:24 +0000 (0:00:03.220)       0:02:02.626 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over nsx.ini] ******************************************",
                                "Tuesday 13 January 2026  01:00:26 +0000 (0:00:02.128)       0:02:04.755 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copy neutron-l3-agent-wrapper script] **************************",
                                "Tuesday 13 January 2026  01:00:28 +0000 (0:00:02.031)       0:02:06.787 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over extra ml2 plugins] ********************************",
                                "Tuesday 13 January 2026  01:00:30 +0000 (0:00:01.902)       0:02:08.689 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************",
                                "Tuesday 13 January 2026  01:00:32 +0000 (0:00:01.826)       0:02:10.515 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_taas.conf] ********************************",
                                "Tuesday 13 January 2026  01:00:35 +0000 (0:00:02.715)       0:02:13.231 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Check neutron containers] **************************************",
                                "Tuesday 13 January 2026  01:00:37 +0000 (0:00:02.798)       0:02:16.030 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:00:40 +0000 (0:00:02.568)       0:02:18.598 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database] *************************************",
                                "Tuesday 13 January 2026  01:00:40 +0000 (0:00:00.458)       0:02:19.056 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database user and setting permissions] ********",
                                "Tuesday 13 January 2026  01:00:42 +0000 (0:00:01.974)       0:02:21.031 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Running Neutron bootstrap container] ***************************",
                                "Tuesday 13 January 2026  01:00:45 +0000 (0:00:02.304)       0:02:23.335 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:24 +0000 (0:00:39.430)       0:03:02.766 ******* ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:24 +0000 (0:00:00.185)       0:03:02.952 ******* ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:25 +0000 (0:00:00.237)       0:03:03.190 ******* ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:25 +0000 (0:00:00.130)       0:03:03.321 ******* ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:25 +0000 (0:00:00.055)       0:03:03.376 ******* ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 13 January 2026  01:01:25 +0000 (0:00:00.099)       0:03:03.476 ******* ",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-server container] *******************",
                                "Tuesday 13 January 2026  01:01:25 +0000 (0:00:00.140)       0:03:03.617 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******",
                                "Tuesday 13 January 2026  01:01:53 +0000 (0:00:28.430)       0:03:32.047 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:02:45 +0000 (0:00:51.904)       0:04:23.951 ******* ",
                                "=============================================================================== ",
                                "neutron : Restart neutron-ovn-metadata-agent container ----------------- 51.90s",
                                "neutron : Running Neutron bootstrap container -------------------------- 39.43s",
                                "neutron : Restart neutron-server container ----------------------------- 28.43s",
                                "service-ks-register : neutron | Granting user roles --------------------- 7.71s",
                                "service-ks-register : neutron | Creating endpoints ---------------------- 7.33s",
                                "neutron : Copying over neutron.conf ------------------------------------- 6.94s",
                                "service-ks-register : neutron | Creating users -------------------------- 4.42s",
                                "neutron : Copying over ml2_conf.ini ------------------------------------- 4.32s",
                                "neutron : Copying over config.json files for services ------------------- 4.24s",
                                "neutron : Ensuring config directories exist ----------------------------- 3.69s",
                                "service-ks-register : neutron | Creating roles -------------------------- 3.68s",
                                "service-ks-register : neutron | Creating services ----------------------- 3.63s",
                                "service-ks-register : neutron | Creating projects ----------------------- 3.51s",
                                "neutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 3.43s",
                                "Setting sysctl values --------------------------------------------------- 3.36s",
                                "neutron : Copying over existing policy file ----------------------------- 3.30s",
                                "neutron : Copying over bgp_dragent.ini ---------------------------------- 3.22s",
                                "service-cert-copy : neutron | Copying over backend internal TLS certificate --- 3.17s",
                                "neutron : Copying over l3_agent.ini ------------------------------------- 2.91s",
                                "neutron : Copying over ssh key ------------------------------------------ 2.90s",
                                "\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463e705-f901-4fd7-827c-1c234e776e5a is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:00:55 +0000 (0:00:00.209)       0:00:00.209 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:00:56 +0000 (0:00:00.596)       0:00:00.806 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m",
                                "",
                                "PLAY [Apply role prometheus] ***************************************************",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  01:00:56 +0000 (0:00:00.602)       0:00:01.408 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [prometheus : Ensuring config directories exist] **************************",
                                "Tuesday 13 January 2026  01:00:58 +0000 (0:00:01.257)       0:00:02.665 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Tuesday 13 January 2026  01:01:00 +0000 (0:00:02.649)       0:00:05.315 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****",
                                "Tuesday 13 January 2026  01:01:02 +0000 (0:00:01.435)       0:00:06.751 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:01:08 +0000 (0:00:06.126)       0:00:12.877 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***",
                                "Tuesday 13 January 2026  01:01:09 +0000 (0:00:01.526)       0:00:14.403 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over config.json files] *****************************",
                                "Tuesday 13 January 2026  01:01:12 +0000 (0:00:02.295)       0:00:16.699 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom prometheus alert rules files] *******************",
                                "Tuesday 13 January 2026  01:01:18 +0000 (0:00:06.763)       0:00:23.463 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom prometheus alert rules files] ***********",
                                "Tuesday 13 January 2026  01:01:20 +0000 (0:00:01.770)       0:00:25.234 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1318086, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.213797, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1318119, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2192214, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1318075, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.21301, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1318105, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2176554, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317917, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1927145, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1318089, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2150676, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1318102, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2172074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1318092, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2158182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1318081, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2137249, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1318115, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2188346, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317912, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192061, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1318133, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2209988, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1318112, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2185857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317921, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2124681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1317913, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.192258, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1318099, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2165592, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1318096, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2159517, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1318131, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.2204368, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus common config overrides] ********************",
                                "Tuesday 13 January 2026  01:01:49 +0000 (0:00:28.699)       0:00:53.933 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus host config overrides] **********************",
                                "Tuesday 13 January 2026  01:01:50 +0000 (0:00:00.720)       0:00:54.654 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus config file] ************************",
                                "Tuesday 13 January 2026  01:01:52 +0000 (0:00:02.029)       0:00:56.684 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus web config file] ********************",
                                "Tuesday 13 January 2026  01:02:11 +0000 (0:00:19.058)       0:01:15.742 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus alertmanager config file] ***********",
                                "Tuesday 13 January 2026  01:02:14 +0000 (0:00:03.602)       0:01:19.345 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom Alertmanager alert notification templates] ******",
                                "Tuesday 13 January 2026  01:02:16 +0000 (0:00:01.424)       0:01:20.769 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom Alertmanager alert notification templates] ***",
                                "Tuesday 13 January 2026  01:02:16 +0000 (0:00:00.620)       0:01:21.390 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************",
                                "Tuesday 13 January 2026  01:02:17 +0000 (0:00:00.580)       0:01:21.970 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying cloud config file for openstack exporter] ***********",
                                "Tuesday 13 January 2026  01:02:19 +0000 (0:00:01.764)       0:01:23.734 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying config file for blackbox exporter] ******************",
                                "Tuesday 13 January 2026  01:02:21 +0000 (0:00:01.857)       0:01:25.592 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Find extra prometheus server config files] ******************",
                                "Tuesday 13 January 2026  01:02:22 +0000 (0:00:01.575)       0:01:27.168 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m",
                                "\u001b[1;35mnot a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Create subdirectories for extra config files] ***************",
                                "Tuesday 13 January 2026  01:02:23 +0000 (0:00:01.011)       0:01:28.179 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Template extra prometheus server config files] **************",
                                "Tuesday 13 January 2026  01:02:24 +0000 (0:00:00.705)       0:01:28.885 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Check prometheus containers] ********************************",
                                "Tuesday 13 January 2026  01:02:25 +0000 (0:00:00.672)       0:01:29.558 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Creating prometheus database user and setting permissions] ***",
                                "Tuesday 13 January 2026  01:02:28 +0000 (0:00:03.618)       0:01:33.177 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:29 +0000 (0:00:01.319)       0:01:34.497 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.081)       0:01:34.578 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.064)       0:01:34.643 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.059)       0:01:34.702 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.225)       0:01:34.928 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.063)       0:01:34.991 ******* ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.063)       0:01:35.055 ******* ",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-server container] *************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.085)       0:01:35.141 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******",
                                "Tuesday 13 January 2026  01:02:47 +0000 (0:00:16.704)       0:01:51.846 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****",
                                "Tuesday 13 January 2026  01:03:01 +0000 (0:00:14.270)       0:02:06.116 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***",
                                "Tuesday 13 January 2026  01:03:11 +0000 (0:00:10.338)       0:02:16.455 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********",
                                "Tuesday 13 January 2026  01:03:24 +0000 (0:00:12.314)       0:02:28.769 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******",
                                "Tuesday 13 January 2026  01:03:38 +0000 (0:00:14.651)       0:02:43.420 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***",
                                "Tuesday 13 January 2026  01:03:47 +0000 (0:00:08.192)       0:02:51.613 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***",
                                "Tuesday 13 January 2026  01:03:51 +0000 (0:00:04.742)       0:02:56.356 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***",
                                "Tuesday 13 January 2026  01:03:56 +0000 (0:00:04.951)       0:03:01.308 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:04:06 +0000 (0:00:09.826)       0:03:11.134 ******* ",
                                "=============================================================================== ",
                                "prometheus : Copying over custom prometheus alert rules files ---------- 28.70s",
                                "prometheus : Copying over prometheus config file ----------------------- 19.06s",
                                "prometheus : Restart prometheus-server container ----------------------- 16.70s",
                                "prometheus : Restart prometheus-cadvisor container --------------------- 14.65s",
                                "prometheus : Restart prometheus-node-exporter container ---------------- 14.27s",
                                "prometheus : Restart prometheus-memcached-exporter container ----------- 12.31s",
                                "prometheus : Restart prometheus-mysqld-exporter container -------------- 10.34s",
                                "prometheus : Restart prometheus-libvirt-exporter container -------------- 9.83s",
                                "prometheus : Restart prometheus-alertmanager container ------------------ 8.19s",
                                "prometheus : Copying over config.json files ----------------------------- 6.76s",
                                "service-cert-copy : prometheus | Copying over extra CA certificates ----- 6.13s",
                                "prometheus : Restart prometheus-blackbox-exporter container ------------- 4.95s",
                                "prometheus : Restart prometheus-elasticsearch-exporter container -------- 4.74s",
                                "prometheus : Check prometheus containers -------------------------------- 3.62s",
                                "prometheus : Copying over prometheus web config file -------------------- 3.60s",
                                "prometheus : Ensuring config directories exist -------------------------- 2.65s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS key --- 2.30s",
                                "prometheus : Find prometheus host config overrides ---------------------- 2.03s",
                                "prometheus : Copying cloud config file for openstack exporter ----------- 1.86s",
                                "prometheus : Find custom prometheus alert rules files ------------------- 1.77s",
                                "\u001b[32m2026-01-13 01:04:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3416391c-81e1-4e37-b2f3-229cd199ed0b is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:02:02 +0000 (0:00:00.245)       0:00:00.245 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:02:03 +0000 (0:00:00.251)       0:00:00.496 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m",
                                "",
                                "PLAY [Apply role glance] *******************************************************",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:03 +0000 (0:00:00.343)       0:00:00.840 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating services] ************************",
                                "Tuesday 13 January 2026  01:02:03 +0000 (0:00:00.446)       0:00:01.286 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating endpoints] ***********************",
                                "Tuesday 13 January 2026  01:02:08 +0000 (0:00:04.079)       0:00:05.365 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating projects] ************************",
                                "Tuesday 13 January 2026  01:02:14 +0000 (0:00:06.159)       0:00:11.525 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating users] ***************************",
                                "Tuesday 13 January 2026  01:02:17 +0000 (0:00:02.823)       0:00:14.348 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating roles] ***************************",
                                "Tuesday 13 January 2026  01:02:20 +0000 (0:00:03.436)       0:00:17.784 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Granting user roles] **********************",
                                "Tuesday 13 January 2026  01:02:23 +0000 (0:00:03.059)       0:00:20.844 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  01:02:26 +0000 (0:00:03.334)       0:00:24.178 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:03.420)       0:00:27.599 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [glance : Ensuring glance service ceph config subdir exists] **************",
                                "Tuesday 13 January 2026  01:02:30 +0000 (0:00:00.663)       0:00:28.262 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copy over multiple ceph configs for Glance] *********************",
                                "Tuesday 13 January 2026  01:02:34 +0000 (0:00:03.638)       0:00:31.901 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Copy over ceph Glance keyrings] *********************************",
                                "Tuesday 13 January 2026  01:02:36 +0000 (0:00:01.656)       0:00:33.557 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directory has correct owner and permission] *****",
                                "Tuesday 13 January 2026  01:02:37 +0000 (0:00:01.292)       0:00:34.850 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check if policies shall be overwritten] *************************",
                                "Tuesday 13 January 2026  01:02:38 +0000 (0:00:00.648)       0:00:35.499 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Set glance policy file] *****************************************",
                                "Tuesday 13 January 2026  01:02:38 +0000 (0:00:00.375)       0:00:35.875 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:38 +0000 (0:00:00.291)       0:00:36.166 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over extra CA certificates] *********",
                                "Tuesday 13 January 2026  01:02:39 +0000 (0:00:00.531)       0:00:36.698 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:02:43 +0000 (0:00:04.540)       0:00:41.239 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS key] ******",
                                "Tuesday 13 January 2026  01:02:48 +0000 (0:00:04.068)       0:00:45.307 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating TLS backend PEM File] **********************************",
                                "Tuesday 13 January 2026  01:02:54 +0000 (0:00:06.092)       0:00:51.399 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  01:02:58 +0000 (0:00:04.250)       0:00:55.650 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-api.conf] ***********************************",
                                "Tuesday 13 January 2026  01:03:02 +0000 (0:00:03.935)       0:00:59.585 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-cache.conf for glance_api] ******************",
                                "Tuesday 13 January 2026  01:03:08 +0000 (0:00:05.897)       0:01:05.482 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-swift.conf for glance_api] ******************",
                                "Tuesday 13 January 2026  01:03:11 +0000 (0:00:03.500)       0:01:08.983 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-image-import.conf] **************************",
                                "Tuesday 13 January 2026  01:03:16 +0000 (0:00:04.952)       0:01:13.935 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over property-protections-rules.conf] *******************",
                                "Tuesday 13 January 2026  01:03:19 +0000 (0:00:03.028)       0:01:16.963 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over existing policy file] ******************************",
                                "Tuesday 13 January 2026  01:03:22 +0000 (0:00:03.011)       0:01:19.974 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-haproxy-tls.cfg] ****************************",
                                "Tuesday 13 January 2026  01:03:22 +0000 (0:00:00.261)       0:01:20.236 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostnqn' file for glance_api] ***********************",
                                "Tuesday 13 January 2026  01:03:29 +0000 (0:00:06.723)       0:01:26.959 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check glance containers] ****************************************",
                                "Tuesday 13 January 2026  01:03:35 +0000 (0:00:05.818)       0:01:32.778 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:03:38 +0000 (0:00:03.476)       0:01:36.255 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database] ***************************************",
                                "Tuesday 13 January 2026  01:03:39 +0000 (0:00:00.254)       0:01:36.509 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database user and setting permissions] **********",
                                "Tuesday 13 January 2026  01:03:41 +0000 (0:00:02.055)       0:01:38.565 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Enable log_bin_trust_function_creators function] ****************",
                                "Tuesday 13 January 2026  01:03:43 +0000 (0:00:02.452)       0:01:41.017 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Running Glance bootstrap container] *****************************",
                                "Tuesday 13 January 2026  01:03:45 +0000 (0:00:02.183)       0:01:43.201 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Disable log_bin_trust_function_creators function] ***************",
                                "Tuesday 13 January 2026  01:04:13 +0000 (0:00:27.604)       0:02:10.805 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:15 +0000 (0:00:02.075)       0:02:12.881 ******* ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:15 +0000 (0:00:00.337)       0:02:13.219 ******* ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:15 +0000 (0:00:00.070)       0:02:13.289 ******* ",
                                "",
                                "RUNNING HANDLER [glance : Restart glance-api container] ************************",
                                "Tuesday 13 January 2026  01:04:16 +0000 (0:00:00.082)       0:02:13.372 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:04:45 +0000 (0:00:29.430)       0:02:42.803 ******* ",
                                "=============================================================================== ",
                                "glance : Restart glance-api container ---------------------------------- 29.43s",
                                "glance : Running Glance bootstrap container ---------------------------- 27.60s",
                                "glance : Copying over glance-haproxy-tls.cfg ---------------------------- 6.72s",
                                "service-ks-register : glance | Creating endpoints ----------------------- 6.16s",
                                "service-cert-copy : glance | Copying over backend internal TLS key ------ 6.09s",
                                "glance : Copying over glance-api.conf ----------------------------------- 5.90s",
                                "glance : Generating 'hostnqn' file for glance_api ----------------------- 5.82s",
                                "glance : Copying over glance-swift.conf for glance_api ------------------ 4.95s",
                                "service-cert-copy : glance | Copying over extra CA certificates --------- 4.54s",
                                "glance : Creating TLS backend PEM File ---------------------------------- 4.25s",
                                "service-ks-register : glance | Creating services ------------------------ 4.08s",
                                "service-cert-copy : glance | Copying over backend internal TLS certificate --- 4.07s",
                                "glance : Copying over config.json files for services -------------------- 3.94s",
                                "glance : Ensuring glance service ceph config subdir exists -------------- 3.64s",
                                "glance : Copying over glance-cache.conf for glance_api ------------------ 3.50s",
                                "glance : Check glance containers ---------------------------------------- 3.48s",
                                "service-ks-register : glance | Creating users --------------------------- 3.44s",
                                "glance : Ensuring config directories exist ------------------------------ 3.42s",
                                "service-ks-register : glance | Granting user roles ---------------------- 3.33s",
                                "service-ks-register : glance | Creating roles --------------------------- 3.06s",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:04:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2e5038ab-5a75-401a-82a0-f3bb852931c1 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:02:23 +0000 (0:00:00.226)       0:00:00.227 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:02:24 +0000 (0:00:00.280)       0:00:00.507 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m",
                                "",
                                "PLAY [Apply role cinder] *******************************************************",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:24 +0000 (0:00:00.346)       0:00:00.853 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating services] ************************",
                                "Tuesday 13 January 2026  01:02:24 +0000 (0:00:00.480)       0:00:01.334 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating endpoints] ***********************",
                                "Tuesday 13 January 2026  01:02:28 +0000 (0:00:03.247)       0:00:04.581 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating projects] ************************",
                                "Tuesday 13 January 2026  01:02:33 +0000 (0:00:05.869)       0:00:10.451 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating users] ***************************",
                                "Tuesday 13 January 2026  01:02:37 +0000 (0:00:03.811)       0:00:14.263 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating roles] ***************************",
                                "Tuesday 13 January 2026  01:02:41 +0000 (0:00:03.702)       0:00:17.965 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Granting user roles] **********************",
                                "Tuesday 13 January 2026  01:02:45 +0000 (0:00:03.850)       0:00:21.816 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directories exist] ******************************",
                                "Tuesday 13 January 2026  01:02:52 +0000 (0:00:07.064)       0:00:28.880 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:55 +0000 (0:00:02.696)       0:00:31.577 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:02:55 +0000 (0:00:00.515)       0:00:32.092 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring cinder service ceph config subdirs exists] *************",
                                "Tuesday 13 January 2026  01:02:56 +0000 (0:00:01.128)       0:00:33.221 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over multiple ceph.conf for cinder services] ************",
                                "Tuesday 13 January 2026  01:02:58 +0000 (0:00:01.896)       0:00:35.117 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************",
                                "Tuesday 13 January 2026  01:03:01 +0000 (0:00:03.239)       0:00:38.357 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************",
                                "Tuesday 13 January 2026  01:03:04 +0000 (0:00:02.436)       0:00:40.794 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directory has correct owner and permission] *****",
                                "Tuesday 13 January 2026  01:03:07 +0000 (0:00:03.127)       0:00:43.921 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Check if policies shall be overwritten] *************************",
                                "Tuesday 13 January 2026  01:03:08 +0000 (0:00:00.976)       0:00:44.898 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Set cinder policy file] *****************************************",
                                "Tuesday 13 January 2026  01:03:08 +0000 (0:00:00.097)       0:00:44.996 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:03:08 +0000 (0:00:00.336)       0:00:45.332 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over extra CA certificates] *********",
                                "Tuesday 13 January 2026  01:03:09 +0000 (0:00:00.646)       0:00:45.979 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:03:14 +0000 (0:00:04.885)       0:00:50.864 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******",
                                "Tuesday 13 January 2026  01:03:15 +0000 (0:00:01.128)       0:00:51.993 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over config.json files for services] ********************",
                                "Tuesday 13 January 2026  01:03:16 +0000 (0:00:01.347)       0:00:53.341 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder-wsgi.conf] **********************************",
                                "Tuesday 13 January 2026  01:03:20 +0000 (0:00:04.119)       0:00:57.460 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder.conf] ***************************************",
                                "Tuesday 13 January 2026  01:03:22 +0000 (0:00:01.838)       0:00:59.299 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************",
                                "Tuesday 13 January 2026  01:03:38 +0000 (0:00:15.185)       0:01:14.485 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over existing policy file] ******************************",
                                "Tuesday 13 January 2026  01:03:39 +0000 (0:00:01.468)       0:01:15.953 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over nfs_shares files for cinder_volume] ****************",
                                "Tuesday 13 January 2026  01:03:40 +0000 (0:00:00.582)       0:01:16.536 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Check cinder containers] ****************************************",
                                "Tuesday 13 January 2026  01:03:40 +0000 (0:00:00.289)       0:01:16.825 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 13 January 2026  01:03:43 +0000 (0:00:03.248)       0:01:20.073 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database] ***************************************",
                                "Tuesday 13 January 2026  01:03:44 +0000 (0:00:00.697)       0:01:20.771 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database user and setting permissions] **********",
                                "Tuesday 13 January 2026  01:03:46 +0000 (0:00:02.178)       0:01:22.950 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Running Cinder bootstrap container] *****************************",
                                "Tuesday 13 January 2026  01:03:48 +0000 (0:00:02.224)       0:01:25.175 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:08 +0000 (0:00:19.633)       0:01:44.808 ******* ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:08 +0000 (0:00:00.067)       0:01:44.876 ******* ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 13 January 2026  01:04:08 +0000 (0:00:00.065)       0:01:44.941 ******* ",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-api container] ************************",
                                "Tuesday 13 January 2026  01:04:08 +0000 (0:00:00.064)       0:01:45.005 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************",
                                "Tuesday 13 January 2026  01:04:32 +0000 (0:00:24.405)       0:02:09.411 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-volume container] *********************",
                                "Tuesday 13 January 2026  01:04:43 +0000 (0:00:10.515)       0:02:19.927 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-backup container] *********************",
                                "Tuesday 13 January 2026  01:05:01 +0000 (0:00:17.661)       0:02:37.588 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***",
                                "Tuesday 13 January 2026  01:05:06 +0000 (0:00:05.728)       0:02:43.317 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:05:07 +0000 (0:00:00.259)       0:02:43.577 ******* ",
                                "=============================================================================== ",
                                "cinder : Restart cinder-api container ---------------------------------- 24.41s",
                                "cinder : Running Cinder bootstrap container ---------------------------- 19.63s",
                                "cinder : Restart cinder-volume container ------------------------------- 17.66s",
                                "cinder : Copying over cinder.conf -------------------------------------- 15.19s",
                                "cinder : Restart cinder-scheduler container ---------------------------- 10.52s",
                                "service-ks-register : cinder | Granting user roles ---------------------- 7.06s",
                                "service-ks-register : cinder | Creating endpoints ----------------------- 5.87s",
                                "cinder : Restart cinder-backup container -------------------------------- 5.73s",
                                "service-cert-copy : cinder | Copying over extra CA certificates --------- 4.89s",
                                "cinder : Copying over config.json files for services -------------------- 4.12s",
                                "service-ks-register : cinder | Creating roles --------------------------- 3.85s",
                                "service-ks-register : cinder | Creating projects ------------------------ 3.81s",
                                "service-ks-register : cinder | Creating users --------------------------- 3.70s",
                                "cinder : Check cinder containers ---------------------------------------- 3.25s",
                                "service-ks-register : cinder | Creating services ------------------------ 3.25s",
                                "cinder : Copying over multiple ceph.conf for cinder services ------------ 3.24s",
                                "cinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.13s",
                                "cinder : Ensuring config directories exist ------------------------------ 2.70s",
                                "cinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.44s",
                                "cinder : Creating Cinder database user and setting permissions ---------- 2.22s",
                                "\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:05:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44d3327a-5e93-4172-95e8-8f720dfb7d65 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3493f45-d74b-4bf5-afa0-2506ace92edf is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:04:11 +0000 (0:00:00.184)       0:00:00.185 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:04:11 +0000 (0:00:00.347)       0:00:00.532 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Nova service] ***********************************************",
                                "",
                                "TASK [Waiting for Nova public port to be UP] ***********************************",
                                "Tuesday 13 January 2026  01:04:12 +0000 (0:00:01.181)       0:00:01.714 ******* ",
                                "",
                                "STILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:06:49 +0000 (0:02:36.889)       0:02:38.604 ******* ",
                                "=============================================================================== ",
                                "Waiting for Nova public port to be UP --------------------------------- 156.89s",
                                "Group hosts based on enabled services ----------------------------------- 1.18s",
                                "Group hosts based on Kolla action --------------------------------------- 0.35s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:04:50 +0000 (0:00:00.275)       0:00:00.275 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:04:50 +0000 (0:00:00.318)       0:00:00.594 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m",
                                "",
                                "PLAY [Apply role grafana] ******************************************************",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:04:51 +0000 (0:00:00.443)       0:00:01.038 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [grafana : Ensuring config directories exist] *****************************",
                                "Tuesday 13 January 2026  01:04:51 +0000 (0:00:00.487)       0:00:01.525 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Check if extra configuration file exists] **********************",
                                "Tuesday 13 January 2026  01:04:52 +0000 (0:00:00.704)       0:00:02.230 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:04:53 +0000 (0:00:00.824)       0:00:03.055 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  01:04:54 +0000 (0:00:00.678)       0:00:03.733 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:04:55 +0000 (0:00:01.341)       0:00:05.075 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  01:04:55 +0000 (0:00:00.374)       0:00:05.450 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over config.json files] ********************************",
                                "Tuesday 13 January 2026  01:04:56 +0000 (0:00:00.823)       0:00:06.273 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over grafana.ini] **************************************",
                                "Tuesday 13 January 2026  01:04:57 +0000 (0:00:01.107)       0:00:07.380 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over extra configuration file] *************************",
                                "Tuesday 13 January 2026  01:04:58 +0000 (0:00:01.135)       0:00:08.515 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Configuring Prometheus as data source for Grafana] *************",
                                "Tuesday 13 January 2026  01:04:59 +0000 (0:00:00.515)       0:00:09.031 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "",
                                "TASK [grafana : Configuring dashboards provisioning] ***************************",
                                "Tuesday 13 January 2026  01:05:00 +0000 (0:00:01.200)       0:00:10.231 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "",
                                "TASK [grafana : Find custom grafana dashboards] ********************************",
                                "Tuesday 13 January 2026  01:05:01 +0000 (0:00:01.187)       0:00:11.419 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : Find templated grafana dashboards] *****************************",
                                "Tuesday 13 January 2026  01:05:02 +0000 (0:00:00.762)       0:00:12.181 ******* ",
                                "\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Prune templated Grafana dashboards] ****************************",
                                "Tuesday 13 January 2026  01:05:03 +0000 (0:00:00.729)       0:00:12.910 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over custom dashboards] ********************************",
                                "Tuesday 13 January 2026  01:05:03 +0000 (0:00:00.516)       0:00:13.427 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1317504, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1011682, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1317572, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1126423, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1317521, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037252, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1317573, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1154642, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1317539, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1317557, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1113179, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1317503, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.0994604, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1317514, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1017954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1317525, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1037955, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1317546, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.108681, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1317571, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1117954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1317518, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1027954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1317553, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.110036, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1317541, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.107623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1317534, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1061764, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1317532, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1048384, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1317551, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1087954, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1317527, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1044922, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1317568, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.111564, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317903, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1910748, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317636, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327233, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1317614, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1317661, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.139, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1317589, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1173158, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317855, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1827967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317669, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.143796, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1317859, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1837895, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317890, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1897967, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1317854, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1819885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317652, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1344657, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1317625, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.125227, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317649, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1327958, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1317616, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1237957, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1317655, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1356792, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317880, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1887968, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317870, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.185888, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1317597, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1182532, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1317605, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1209862, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317682, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.1460068, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1317863, 'dev': 125, 'nlink': 1, 'atime': 1768262556.0, 'mtime': 1768262556.0, 'ctime': 1768263336.184734, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "",
                                "TASK [grafana : Check grafana containers] **************************************",
                                "Tuesday 13 January 2026  01:05:39 +0000 (0:00:35.874)       0:00:49.302 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database] *************************************",
                                "Tuesday 13 January 2026  01:05:40 +0000 (0:00:01.050)       0:00:50.352 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database user and setting permissions] ********",
                                "Tuesday 13 January 2026  01:05:42 +0000 (0:00:02.302)       0:00:52.654 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:05:45 +0000 (0:00:02.216)       0:00:54.871 ******* ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:05:45 +0000 (0:00:00.063)       0:00:54.934 ******* ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:05:45 +0000 (0:00:00.060)       0:00:54.994 ******* ",
                                "",
                                "RUNNING HANDLER [grafana : Restart first grafana container] ********************",
                                "Tuesday 13 January 2026  01:05:45 +0000 (0:00:00.225)       0:00:55.219 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********",
                                "Tuesday 13 January 2026  01:05:47 +0000 (0:00:01.858)       0:00:57.078 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Restart remaining grafana containers] ***************",
                                "Tuesday 13 January 2026  01:06:25 +0000 (0:00:38.331)       0:01:35.409 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [grafana : Wait for grafana application ready] ****************************",
                                "Tuesday 13 January 2026  01:06:52 +0000 (0:00:26.777)       0:02:02.187 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Remove old grafana docker volume] ******************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:01.950)       0:02:04.138 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Enable grafana datasources] ************************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:00.361)       0:02:04.499 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Disable Getting Started panel] *********************************",
                                "Tuesday 13 January 2026  01:06:57 +0000 (0:00:02.545)       0:02:07.045 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:06:57 +0000 (0:00:00.244)       0:02:07.289 ******* ",
                                "=============================================================================== ",
                                "grafana : Waiting for grafana to start on first node ------------------- 38.33s",
                                "grafana : Copying over custom dashboards ------------------------------- 35.87s",
                                "grafana : Restart remaining grafana containers ------------------------- 26.78s",
                                "grafana : Enable grafana datasources ------------------------------------ 2.55s",
                                "grafana : Creating grafana database ------------------------------------- 2.30s",
                                "grafana : Creating grafana database user and setting permissions -------- 2.22s",
                                "grafana : Wait for grafana application ready ---------------------------- 1.95s",
                                "grafana : Restart first grafana container ------------------------------- 1.86s",
                                "service-cert-copy : grafana | Copying over extra CA certificates -------- 1.34s",
                                "grafana : Configuring Prometheus as data source for Grafana ------------- 1.20s",
                                "grafana : Configuring dashboards provisioning --------------------------- 1.19s",
                                "grafana : Copying over grafana.ini -------------------------------------- 1.14s",
                                "grafana : Copying over config.json files -------------------------------- 1.11s",
                                "grafana : Check grafana containers -------------------------------------- 1.05s",
                                "grafana : Check if extra configuration file exists ---------------------- 0.82s",
                                "service-cert-copy : grafana | Copying over backend internal TLS key ----- 0.82s",
                                "grafana : Find custom grafana dashboards -------------------------------- 0.76s",
                                "grafana : Find templated grafana dashboards ----------------------------- 0.73s",
                                "grafana : Ensuring config directories exist ----------------------------- 0.70s",
                                "grafana : include_tasks ------------------------------------------------- 0.68s",
                                "\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:06:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:07:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:08:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:09:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 40df7405-e8dc-40ca-8ff8-36b16a2c0c2a is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on OpenStack release] **********************************",
                                "Tuesday 13 January 2026  01:02:54 +0000 (0:00:00.481)       0:00:00.481 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:02:56 +0000 (0:00:01.503)       0:00:01.985 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:02:56 +0000 (0:00:00.858)       0:00:02.843 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova API databases] ********************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Tuesday 13 January 2026  01:02:57 +0000 (0:00:00.933)       0:00:03.777 ******* ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases] ******************************************",
                                "Tuesday 13 January 2026  01:02:58 +0000 (0:00:00.737)       0:00:04.514 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases user and setting permissions] *************",
                                "Tuesday 13 January 2026  01:03:02 +0000 (0:00:03.999)       0:00:08.514 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Tuesday 13 January 2026  01:03:07 +0000 (0:00:04.448)       0:00:12.963 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for nova-api-bootstrap] ************",
                                "Tuesday 13 January 2026  01:03:07 +0000 (0:00:00.589)       0:00:13.552 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************",
                                "Tuesday 13 January 2026  01:03:08 +0000 (0:00:01.133)       0:00:14.686 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 13 January 2026  01:03:11 +0000 (0:00:02.835)       0:00:17.521 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Tuesday 13 January 2026  01:03:12 +0000 (0:00:00.344)       0:00:17.866 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Create cell0 mappings] ********************************************",
                                "Tuesday 13 January 2026  01:03:43 +0000 (0:00:31.023)       0:00:48.889 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 13 January 2026  01:03:58 +0000 (0:00:15.495)       0:01:04.384 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 13 January 2026  01:04:10 +0000 (0:00:11.758)       0:01:16.143 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Update cell0 mappings] ********************************************",
                                "Tuesday 13 January 2026  01:04:11 +0000 (0:00:01.248)       0:01:17.392 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 13 January 2026  01:04:12 +0000 (0:00:00.530)       0:01:17.922 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Tuesday 13 January 2026  01:04:12 +0000 (0:00:00.488)       0:01:18.411 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Tuesday 13 January 2026  01:04:30 +0000 (0:00:18.251)       0:01:36.663 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova cell databases] *******************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Tuesday 13 January 2026  01:04:31 +0000 (0:00:00.249)       0:01:36.913 ******* ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database] *********************************",
                                "Tuesday 13 January 2026  01:04:31 +0000 (0:00:00.442)       0:01:37.355 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database user and setting permissions] ****",
                                "Tuesday 13 January 2026  01:04:33 +0000 (0:00:02.014)       0:01:39.370 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Tuesday 13 January 2026  01:04:35 +0000 (0:00:02.284)       0:01:41.654 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Tuesday 13 January 2026  01:04:36 +0000 (0:00:00.361)       0:01:42.016 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Tuesday 13 January 2026  01:04:43 +0000 (0:00:06.992)       0:01:49.009 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Tuesday 13 January 2026  01:04:43 +0000 (0:00:00.321)       0:01:49.331 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Tuesday 13 January 2026  01:04:44 +0000 (0:00:00.972)       0:01:50.303 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******",
                                "Tuesday 13 January 2026  01:04:45 +0000 (0:00:00.817)       0:01:51.121 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************",
                                "Tuesday 13 January 2026  01:04:46 +0000 (0:00:00.944)       0:01:52.065 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Running Nova cell bootstrap container] ***********************",
                                "Tuesday 13 January 2026  01:04:48 +0000 (0:00:02.147)       0:01:54.213 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 13 January 2026  01:05:09 +0000 (0:00:20.709)       0:02:14.922 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 13 January 2026  01:05:21 +0000 (0:00:12.434)       0:02:27.356 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Create cell] *************************************************",
                                "Tuesday 13 January 2026  01:05:22 +0000 (0:00:00.840)       0:02:28.197 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Update cell] *************************************************",
                                "Tuesday 13 January 2026  01:05:36 +0000 (0:00:13.770)       0:02:41.968 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Tuesday 13 January 2026  01:05:37 +0000 (0:00:01.125)       0:02:43.093 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova] *********************************************************",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 13 January 2026  01:05:37 +0000 (0:00:00.583)       0:02:43.677 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating services] **************************",
                                "Tuesday 13 January 2026  01:05:38 +0000 (0:00:00.550)       0:02:44.227 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating endpoints] *************************",
                                "Tuesday 13 January 2026  01:05:41 +0000 (0:00:03.403)       0:02:47.630 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating projects] **************************",
                                "Tuesday 13 January 2026  01:05:48 +0000 (0:00:06.635)       0:02:54.266 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating users] *****************************",
                                "Tuesday 13 January 2026  01:05:51 +0000 (0:00:02.940)       0:02:57.206 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating roles] *****************************",
                                "Tuesday 13 January 2026  01:05:55 +0000 (0:00:03.892)       0:03:01.099 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Granting user roles] ************************",
                                "Tuesday 13 January 2026  01:05:58 +0000 (0:00:03.096)       0:03:04.195 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Tuesday 13 January 2026  01:06:05 +0000 (0:00:07.283)       0:03:11.479 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Check if policies shall be overwritten] ***************************",
                                "Tuesday 13 January 2026  01:06:07 +0000 (0:00:01.439)       0:03:12.918 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Set nova policy file] *********************************************",
                                "Tuesday 13 January 2026  01:06:07 +0000 (0:00:00.132)       0:03:13.051 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check for vendordata file] ****************************************",
                                "Tuesday 13 January 2026  01:06:07 +0000 (0:00:00.279)       0:03:13.331 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova : Set vendordata file path] *****************************************",
                                "Tuesday 13 January 2026  01:06:08 +0000 (0:00:00.860)       0:03:14.191 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 13 January 2026  01:06:08 +0000 (0:00:00.275)       0:03:14.467 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Tuesday 13 January 2026  01:06:09 +0000 (0:00:00.528)       0:03:14.995 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:06:11 +0000 (0:00:02.601)       0:03:17.596 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Tuesday 13 January 2026  01:06:12 +0000 (0:00:00.601)       0:03:18.198 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for services] **********************",
                                "Tuesday 13 January 2026  01:06:13 +0000 (0:00:00.806)       0:03:19.005 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf] *******************************************",
                                "Tuesday 13 January 2026  01:06:16 +0000 (0:00:02.911)       0:03:21.916 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over existing policy file] ********************************",
                                "Tuesday 13 January 2026  01:06:21 +0000 (0:00:05.339)       0:03:27.255 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-api-wsgi.conf] **********************************",
                                "Tuesday 13 January 2026  01:06:21 +0000 (0:00:00.572)       0:03:27.828 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over vendordata file] *************************************",
                                "Tuesday 13 January 2026  01:06:23 +0000 (0:00:01.422)       0:03:29.250 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check nova containers] ********************************************",
                                "Tuesday 13 January 2026  01:06:23 +0000 (0:00:00.329)       0:03:29.580 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 13 January 2026  01:06:25 +0000 (0:00:02.118)       0:03:31.699 ******* ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 13 January 2026  01:06:25 +0000 (0:00:00.139)       0:03:31.839 ******* ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 13 January 2026  01:06:26 +0000 (0:00:00.135)       0:03:31.974 ******* ",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-scheduler container] **********************",
                                "Tuesday 13 January 2026  01:06:26 +0000 (0:00:00.152)       0:03:32.126 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-api container] ****************************",
                                "Tuesday 13 January 2026  01:06:41 +0000 (0:00:15.330)       0:03:47.456 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role nova-cell] ****************************************************",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:06:52 +0000 (0:00:11.102)       0:03:58.559 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:01.376)       0:03:59.935 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Load and persist br_netfilter module] ************************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:00.523)       0:04:00.459 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 13 January 2026  01:06:55 +0000 (0:00:00.901)       0:04:01.360 ******* ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 13 January 2026  01:06:56 +0000 (0:00:00.602)       0:04:01.963 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 13 January 2026  01:06:57 +0000 (0:00:01.192)       0:04:03.155 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************",
                                "Tuesday 13 January 2026  01:06:57 +0000 (0:00:00.443)       0:04:03.599 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "",
                                "TASK [nova-cell : Install udev kolla kvm rules] ********************************",
                                "Tuesday 13 January 2026  01:06:58 +0000 (0:00:01.225)       0:04:04.824 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Mask qemu-kvm service] ***************************************",
                                "Tuesday 13 January 2026  01:07:00 +0000 (0:00:01.134)       0:04:05.958 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Tuesday 13 January 2026  01:07:01 +0000 (0:00:01.787)       0:04:07.746 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:07:03 +0000 (0:00:02.011)       0:04:09.758 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Tuesday 13 January 2026  01:07:05 +0000 (0:00:01.374)       0:04:11.132 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:07:08 +0000 (0:00:03.150)       0:04:14.283 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Tuesday 13 January 2026  01:07:10 +0000 (0:00:01.570)       0:04:15.854 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:07:12 +0000 (0:00:02.261)       0:04:18.115 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova keyring file] *************************************",
                                "Tuesday 13 January 2026  01:07:13 +0000 (0:00:01.053)       0:04:19.169 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check cinder keyring file] ***********************************",
                                "Tuesday 13 January 2026  01:07:14 +0000 (0:00:00.946)       0:04:20.115 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract nova key from file] **********************************",
                                "Tuesday 13 January 2026  01:07:15 +0000 (0:00:00.874)       0:04:20.990 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract cinder key from file] ********************************",
                                "Tuesday 13 January 2026  01:07:15 +0000 (0:00:00.486)       0:04:21.476 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph nova keyring file] ****************************",
                                "Tuesday 13 January 2026  01:07:16 +0000 (0:00:00.759)       0:04:22.236 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph cinder keyring file] **************************",
                                "Tuesday 13 January 2026  01:07:17 +0000 (0:00:01.050)       0:04:23.287 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf] *****************************************",
                                "Tuesday 13 January 2026  01:07:18 +0000 (0:00:01.071)       0:04:24.358 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************",
                                "Tuesday 13 January 2026  01:07:21 +0000 (0:00:03.291)       0:04:27.650 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************",
                                "Tuesday 13 January 2026  01:07:22 +0000 (0:00:00.545)       0:04:28.195 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring libvirt secrets directory exists] *******************",
                                "Tuesday 13 January 2026  01:07:22 +0000 (0:00:00.310)       0:04:28.506 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing nova secret xml for libvirt] *************************",
                                "Tuesday 13 January 2026  01:07:23 +0000 (0:00:01.119)       0:04:29.625 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing secrets key for libvirt] *****************************",
                                "Tuesday 13 January 2026  01:07:26 +0000 (0:00:03.072)       0:04:32.697 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check if policies shall be overwritten] **********************",
                                "Tuesday 13 January 2026  01:07:29 +0000 (0:00:02.996)       0:04:35.694 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set nova policy file] ****************************************",
                                "Tuesday 13 January 2026  01:07:29 +0000 (0:00:00.126)       0:04:35.821 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check for vendordata file] ***********************************",
                                "Tuesday 13 January 2026  01:07:30 +0000 (0:00:00.594)       0:04:36.416 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set vendordata file path] ************************************",
                                "Tuesday 13 January 2026  01:07:31 +0000 (0:00:00.676)       0:04:37.092 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for services] *****************",
                                "Tuesday 13 January 2026  01:07:32 +0000 (0:00:00.833)       0:04:37.925 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf] **************************************",
                                "Tuesday 13 January 2026  01:07:35 +0000 (0:00:03.308)       0:04:41.234 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over Nova compute provider config] *******************",
                                "Tuesday 13 January 2026  01:07:41 +0000 (0:00:06.347)       0:04:47.582 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt configuration] **************************",
                                "Tuesday 13 January 2026  01:07:43 +0000 (0:00:01.393)       0:04:48.975 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt TLS keys] *******************************",
                                "Tuesday 13 January 2026  01:07:46 +0000 (0:00:03.656)       0:04:52.631 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt SASL configuration] *********************",
                                "Tuesday 13 January 2026  01:07:47 +0000 (0:00:00.613)       0:04:53.245 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying files for nova-ssh] **********************************",
                                "Tuesday 13 January 2026  01:07:52 +0000 (0:00:05.086)       0:04:58.332 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying VMware vCenter CA file] ******************************",
                                "Tuesday 13 January 2026  01:07:58 +0000 (0:00:06.504)       0:05:04.836 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying 'release' file for nova_compute] *********************",
                                "Tuesday 13 January 2026  01:07:59 +0000 (0:00:00.796)       0:05:05.633 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************",
                                "Tuesday 13 January 2026  01:08:00 +0000 (0:00:00.575)       0:05:06.209 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over existing policy file] ***************************",
                                "Tuesday 13 January 2026  01:08:02 +0000 (0:00:02.045)       0:05:08.254 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over vendordata file to containers] ******************",
                                "Tuesday 13 January 2026  01:08:03 +0000 (0:00:01.211)       0:05:09.466 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova-cell containers] **********************************",
                                "Tuesday 13 January 2026  01:08:04 +0000 (0:00:00.834)       0:05:10.300 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 13 January 2026  01:08:07 +0000 (0:00:03.009)       0:05:13.309 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.745)       0:05:14.055 ******* ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.134)       0:05:14.189 ******* ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.130)       0:05:14.320 ******* ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.135)       0:05:14.456 ******* ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.137)       0:05:14.594 ******* ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 13 January 2026  01:08:08 +0000 (0:00:00.129)       0:05:14.723 ******* ",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************",
                                "Tuesday 13 January 2026  01:08:09 +0000 (0:00:00.362)       0:05:15.086 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************",
                                "Tuesday 13 January 2026  01:08:18 +0000 (0:00:09.493)       0:05:24.580 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************",
                                "Tuesday 13 January 2026  01:08:29 +0000 (0:00:10.888)       0:05:35.468 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************",
                                "Tuesday 13 January 2026  01:08:44 +0000 (0:00:15.133)       0:05:50.602 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************",
                                "Tuesday 13 January 2026  01:09:09 +0000 (0:00:24.319)       0:06:14.922 ******* ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************",
                                "Tuesday 13 January 2026  01:09:15 +0000 (0:00:06.168)       0:06:21.090 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-compute container] *******************",
                                "Tuesday 13 January 2026  01:09:15 +0000 (0:00:00.751)       0:06:21.842 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***",
                                "Tuesday 13 January 2026  01:09:38 +0000 (0:00:22.420)       0:06:44.262 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Waiting for nova-compute services to register themselves] ****",
                                "Tuesday 13 January 2026  01:09:38 +0000 (0:00:00.123)       0:06:44.386 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if nova-compute service failed to register] *************",
                                "Tuesday 13 January 2026  01:09:59 +0000 (0:00:20.923)       0:07:05.309 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include discover_computes.yml] *******************************",
                                "Tuesday 13 January 2026  01:10:07 +0000 (0:00:07.722)       0:07:13.031 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 13 January 2026  01:10:10 +0000 (0:00:03.231)       0:07:16.263 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 13 January 2026  01:10:24 +0000 (0:00:13.846)       0:07:30.109 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if cell settings not found] *****************************",
                                "Tuesday 13 January 2026  01:10:25 +0000 (0:00:01.234)       0:07:31.344 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Discover nova hosts] *****************************************",
                                "Tuesday 13 January 2026  01:10:26 +0000 (0:00:01.284)       0:07:32.629 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************",
                                "Tuesday 13 January 2026  01:10:39 +0000 (0:00:12.681)       0:07:45.310 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Refresh nova scheduler cell cache] ***************************************",
                                "",
                                "TASK [nova : Refresh cell cache in nova scheduler] *****************************",
                                "Tuesday 13 January 2026  01:10:41 +0000 (0:00:01.689)       0:07:46.999 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova super conductor services] *****************************",
                                "",
                                "TASK [nova : Reload nova super conductor services to remove RPC version pin] ***",
                                "Tuesday 13 January 2026  01:10:42 +0000 (0:00:01.134)       0:07:48.134 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload Nova cell services] ***********************************************",
                                "",
                                "TASK [nova-cell : Reload nova cell services to remove RPC version cap] *********",
                                "Tuesday 13 January 2026  01:10:42 +0000 (0:00:00.530)       0:07:48.664 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova API services] *****************************************",
                                "",
                                "TASK [nova : Reload nova API services to remove RPC version pin] ***************",
                                "Tuesday 13 January 2026  01:10:44 +0000 (0:00:01.364)       0:07:50.029 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Run Nova API online data migrations] *************************************",
                                "",
                                "TASK [nova : Run Nova API online database migrations] **************************",
                                "Tuesday 13 January 2026  01:10:44 +0000 (0:00:00.747)       0:07:50.776 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Run Nova cell online data migrations] ************************************",
                                "",
                                "TASK [nova-cell : Run Nova cell online database migrations] ********************",
                                "Tuesday 13 January 2026  01:10:45 +0000 (0:00:00.622)       0:07:51.399 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=44  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=51  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=19  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:10:45 +0000 (0:00:00.409)       0:07:51.808 ******* ",
                                "=============================================================================== ",
                                "nova : Running Nova API bootstrap container ---------------------------- 31.02s",
                                "nova-cell : Restart nova-libvirt container ----------------------------- 24.32s",
                                "nova-cell : Restart nova-compute container ----------------------------- 22.42s",
                                "nova-cell : Waiting for nova-compute services to register themselves --- 20.92s",
                                "nova-cell : Running Nova cell bootstrap container ---------------------- 20.71s",
                                "nova : Running Nova API bootstrap container ---------------------------- 18.25s",
                                "nova : Create cell0 mappings ------------------------------------------- 15.50s",
                                "nova : Restart nova-scheduler container -------------------------------- 15.33s",
                                "nova-cell : Restart nova-ssh container --------------------------------- 15.13s",
                                "nova-cell : Get a list of existing cells ------------------------------- 13.85s",
                                "nova-cell : Create cell ------------------------------------------------ 13.77s",
                                "nova-cell : Discover nova hosts ---------------------------------------- 12.68s",
                                "nova-cell : Get a list of existing cells ------------------------------- 12.43s",
                                "nova-cell : Get a list of existing cells ------------------------------- 11.76s",
                                "nova : Restart nova-api container -------------------------------------- 11.10s",
                                "nova-cell : Restart nova-novncproxy container -------------------------- 10.89s",
                                "nova-cell : Restart nova-conductor container ---------------------------- 9.49s",
                                "nova-cell : Fail if nova-compute service failed to register ------------- 7.72s",
                                "service-ks-register : nova | Granting user roles ------------------------ 7.28s",
                                "service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 6.99s",
                                "\u001b[32m2026-01-13 01:10:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:10:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bcd6be5f-ed3e-4ed4-95a6-2b053d117117 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:00.205)       0:00:00.205 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:00.246)       0:00:00.451 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m",
                                "",
                                "PLAY [Apply role octavia] ******************************************************",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:06:54 +0000 (0:00:00.382)       0:00:00.834 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating services] ***********************",
                                "Tuesday 13 January 2026  01:06:55 +0000 (0:00:00.508)       0:00:01.343 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating endpoints] **********************",
                                "Tuesday 13 January 2026  01:06:58 +0000 (0:00:03.349)       0:00:04.692 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating projects] ***********************",
                                "Tuesday 13 January 2026  01:07:04 +0000 (0:00:06.076)       0:00:10.768 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating users] **************************",
                                "Tuesday 13 January 2026  01:07:07 +0000 (0:00:02.841)       0:00:13.609 ******* ",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating roles] **************************",
                                "Tuesday 13 January 2026  01:07:14 +0000 (0:00:06.613)       0:00:20.223 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Granting user roles] *********************",
                                "Tuesday 13 January 2026  01:07:17 +0000 (0:00:02.887)       0:00:23.110 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "",
                                "TASK [octavia : Adding octavia related roles] **********************************",
                                "Tuesday 13 January 2026  01:07:22 +0000 (0:00:05.787)       0:00:28.898 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:07:36 +0000 (0:00:13.548)       0:00:42.447 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create amphora flavor] *****************************************",
                                "Tuesday 13 January 2026  01:07:37 +0000 (0:00:00.896)       0:00:43.344 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create nova keypair for amphora] *******************************",
                                "Tuesday 13 January 2026  01:07:42 +0000 (0:00:05.143)       0:00:48.488 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Tuesday 13 January 2026  01:07:46 +0000 (0:00:04.121)       0:00:52.609 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create security groups for octavia] ****************************",
                                "Tuesday 13 January 2026  01:07:49 +0000 (0:00:03.151)       0:00:55.761 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Add rules for security groups] *********************************",
                                "Tuesday 13 January 2026  01:07:58 +0000 (0:00:09.166)       0:01:04.927 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management network] ************************",
                                "Tuesday 13 January 2026  01:08:13 +0000 (0:00:14.921)       0:01:19.849 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management subnet] *************************",
                                "Tuesday 13 January 2026  01:08:17 +0000 (0:00:03.755)       0:01:23.604 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management router for IPv6] ****************",
                                "Tuesday 13 January 2026  01:08:22 +0000 (0:00:05.177)       0:01:28.782 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update loadbalancer management subnet] *************************",
                                "Tuesday 13 January 2026  01:08:22 +0000 (0:00:00.224)       0:01:29.007 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:08:27 +0000 (0:00:04.178)       0:01:33.185 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create ports for Octavia health-manager nodes] *****************",
                                "Tuesday 13 January 2026  01:08:28 +0000 (0:00:00.941)       0:01:34.127 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Update Octavia health manager port host_id] ********************",
                                "Tuesday 13 January 2026  01:08:32 +0000 (0:00:04.274)       0:01:38.401 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Add Octavia port to openvswitch br-int] ************************",
                                "Tuesday 13 January 2026  01:08:36 +0000 (0:00:04.455)       0:01:42.856 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Install isc-dhcp-client package] *******************************",
                                "Tuesday 13 January 2026  01:08:37 +0000 (0:00:00.737)       0:01:43.594 ******* ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia dhclient conf] **********************************",
                                "Tuesday 13 January 2026  01:08:39 +0000 (0:00:02.051)       0:01:45.646 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia-interface service] ******************************",
                                "Tuesday 13 January 2026  01:08:40 +0000 (0:00:01.175)       0:01:46.821 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Restart octavia-interface.service if required] *****************",
                                "Tuesday 13 January 2026  01:08:41 +0000 (0:00:01.028)       0:01:47.849 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Enable and start octavia-interface.service] ********************",
                                "Tuesday 13 January 2026  01:08:43 +0000 (0:00:01.793)       0:01:49.643 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Wait for interface ohm0 ip appear] *****************************",
                                "Tuesday 13 January 2026  01:08:45 +0000 (0:00:01.707)       0:01:51.350 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Gather facts] **************************************************",
                                "Tuesday 13 January 2026  01:08:45 +0000 (0:00:00.609)       0:01:51.959 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:08:48 +0000 (0:00:02.448)       0:01:54.408 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Get amphora flavor info] ***************************************",
                                "Tuesday 13 January 2026  01:08:49 +0000 (0:00:00.765)       0:01:55.173 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Tuesday 13 January 2026  01:08:52 +0000 (0:00:03.287)       0:01:58.461 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get security groups for octavia] *******************************",
                                "Tuesday 13 January 2026  01:08:55 +0000 (0:00:03.250)       0:02:01.711 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Get loadbalancer management network] ***************************",
                                "Tuesday 13 January 2026  01:09:01 +0000 (0:00:05.803)       0:02:07.515 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia resources facts] ***********************************",
                                "Tuesday 13 January 2026  01:09:04 +0000 (0:00:02.819)       0:02:10.335 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Ensuring config directories exist] *****************************",
                                "Tuesday 13 January 2026  01:09:04 +0000 (0:00:00.299)       0:02:10.635 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Check if policies shall be overwritten] ************************",
                                "Tuesday 13 January 2026  01:09:06 +0000 (0:00:02.065)       0:02:12.700 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia policy file] ***************************************",
                                "Tuesday 13 January 2026  01:09:06 +0000 (0:00:00.139)       0:02:12.839 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over existing policy file] *****************************",
                                "Tuesday 13 January 2026  01:09:07 +0000 (0:00:00.481)       0:02:13.321 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:09:07 +0000 (0:00:00.691)       0:02:14.013 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over extra CA certificates] ********",
                                "Tuesday 13 January 2026  01:09:08 +0000 (0:00:00.543)       0:02:14.556 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***",
                                "Tuesday 13 January 2026  01:09:13 +0000 (0:00:04.758)       0:02:19.314 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****",
                                "Tuesday 13 January 2026  01:09:14 +0000 (0:00:00.955)       0:02:20.270 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over config.json files for services] *******************",
                                "Tuesday 13 January 2026  01:09:15 +0000 (0:00:00.844)       0:02:21.115 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia-wsgi.conf] ********************************",
                                "Tuesday 13 January 2026  01:09:19 +0000 (0:00:04.640)       0:02:25.755 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia.conf] *************************************",
                                "Tuesday 13 January 2026  01:09:22 +0000 (0:00:02.388)       0:02:28.143 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over Octavia SSH key] **********************************",
                                "Tuesday 13 January 2026  01:09:38 +0000 (0:00:16.314)       0:02:44.457 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-worker] ******************",
                                "Tuesday 13 January 2026  01:09:39 +0000 (0:00:01.495)       0:02:45.953 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-housekeeping] ************",
                                "Tuesday 13 January 2026  01:09:46 +0000 (0:00:06.086)       0:02:52.039 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-health-manager] **********",
                                "Tuesday 13 January 2026  01:09:51 +0000 (0:00:05.063)       0:02:57.103 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Check octavia containers] **************************************",
                                "Tuesday 13 January 2026  01:09:55 +0000 (0:00:04.638)       0:03:01.741 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:09:59 +0000 (0:00:03.506)       0:03:05.248 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database] *************************************",
                                "Tuesday 13 January 2026  01:09:59 +0000 (0:00:00.297)       0:03:05.545 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database] *************************",
                                "Tuesday 13 January 2026  01:10:01 +0000 (0:00:01.830)       0:03:07.375 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database user and setting permissions] ********",
                                "Tuesday 13 January 2026  01:10:03 +0000 (0:00:01.911)       0:03:09.286 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database user and setting permissions] ***",
                                "Tuesday 13 January 2026  01:10:05 +0000 (0:00:02.219)       0:03:11.506 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Running Octavia bootstrap container] ***************************",
                                "Tuesday 13 January 2026  01:10:08 +0000 (0:00:02.550)       0:03:14.057 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:10:30 +0000 (0:00:22.676)       0:03:36.733 ******* ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:10:30 +0000 (0:00:00.064)       0:03:36.797 ******* ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 13 January 2026  01:10:30 +0000 (0:00:00.068)       0:03:36.866 ******* ",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-api container] **********************",
                                "Tuesday 13 January 2026  01:10:30 +0000 (0:00:00.078)       0:03:36.944 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************",
                                "Tuesday 13 January 2026  01:10:46 +0000 (0:00:15.464)       0:03:52.409 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********",
                                "Tuesday 13 January 2026  01:10:52 +0000 (0:00:06.024)       0:03:58.433 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************",
                                "Tuesday 13 January 2026  01:11:02 +0000 (0:00:10.195)       0:04:08.629 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-worker container] *******************",
                                "Tuesday 13 January 2026  01:11:10 +0000 (0:00:08.123)       0:04:16.752 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:11:16 +0000 (0:00:05.519)       0:04:22.272 ******* ",
                                "=============================================================================== ",
                                "octavia : Running Octavia bootstrap container -------------------------- 22.68s",
                                "octavia : Copying over octavia.conf ------------------------------------ 16.31s",
                                "octavia : Restart octavia-api container -------------------------------- 15.46s",
                                "octavia : Add rules for security groups -------------------------------- 14.92s",
                                "octavia : Adding octavia related roles --------------------------------- 13.55s",
                                "octavia : Restart octavia-health-manager container --------------------- 10.20s",
                                "octavia : Create security groups for octavia ---------------------------- 9.17s",
                                "octavia : Restart octavia-housekeeping container ------------------------ 8.12s",
                                "service-ks-register : octavia | Creating users -------------------------- 6.61s",
                                "octavia : Copying certificate files for octavia-worker ------------------ 6.09s",
                                "service-ks-register : octavia | Creating endpoints ---------------------- 6.08s",
                                "octavia : Restart octavia-driver-agent container ------------------------ 6.02s",
                                "octavia : Get security groups for octavia ------------------------------- 5.80s",
                                "service-ks-register : octavia | Granting user roles --------------------- 5.79s",
                                "octavia : Restart octavia-worker container ------------------------------ 5.52s",
                                "octavia : Create loadbalancer management subnet ------------------------- 5.18s",
                                "octavia : Create amphora flavor ----------------------------------------- 5.14s",
                                "octavia : Copying certificate files for octavia-housekeeping ------------ 5.06s",
                                "service-cert-copy : octavia | Copying over extra CA certificates -------- 4.76s",
                                "octavia : Copying over config.json files for services ------------------- 4.64s",
                                "\u001b[32m2026-01-13 01:11:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:11:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- END -- Tue Jan 13 01:12:19 UTC 2026"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001c-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:12:20.494591Z",
                            "start": "2026-01-13T00:38:31.108846Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001c",
                        "name": "Deploy in a nutshell"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "/opt/configuration/scripts/bootstrap.sh"
                            ],
                            "delta": "0:03:06.008900",
                            "end": "2026-01-13 01:15:27.090100",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 /opt/configuration/scripts/bootstrap.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001d-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 01:12:21.081200",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# BOOTSTRAP\n\n+ set -e\n+ echo\n+ echo '# BOOTSTRAP'\n+ echo\n+ sh -c /opt/configuration/scripts/bootstrap-services.sh\n+ set -e\n+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh\n\u001b[32m2026-01-13 01:12:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 65a2231a-56ff-4016-a60d-b8b13fdafb9c (flavor-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-01-13 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m\n\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m\n\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m\n\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m\n\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m\n\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m\n\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m\n\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m\n\u001b[32m2026-01-13 01:12:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m\n\u001b[32m2026-01-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c177d464-cf3b-4607-b28c-f60cc9756e07 (bootstrap-basic) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c177d464-cf3b-4607-b28c-f60cc9756e07 (bootstrap-basic) has been started and output is visible here.\u001b[0m\n\nPLAY [Bootstrap basic OpenStack services] **************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 13 January 2026  01:12:49 +0000 (0:00:00.065)       0:00:00.065 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Get volume type LUKS] ****************************************************\nTuesday 13 January 2026  01:12:51 +0000 (0:00:01.806)       0:00:01.872 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Create volume type LUKS] *************************************************\nTuesday 13 January 2026  01:12:59 +0000 (0:00:08.537)       0:00:10.410 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public network] ***************************************************\nTuesday 13 January 2026  01:13:07 +0000 (0:00:07.991)       0:00:18.402 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Set public network to default] *******************************************\nTuesday 13 January 2026  01:13:13 +0000 (0:00:05.944)       0:00:24.346 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public subnet] ****************************************************\nTuesday 13 January 2026  01:13:19 +0000 (0:00:06.144)       0:00:30.490 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create default IPv4 subnet pool] *****************************************\nTuesday 13 January 2026  01:13:23 +0000 (0:00:04.015)       0:00:34.506 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create manager role] *****************************************************\nTuesday 13 January 2026  01:13:27 +0000 (0:00:03.598)       0:00:38.105 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:13:30 +0000 (0:00:03.376)       0:00:41.482 ******* \n=============================================================================== \nGet volume type LUKS ---------------------------------------------------- 8.54s\nCreate volume type LUKS ------------------------------------------------- 7.99s\nSet public network to default ------------------------------------------- 6.14s\nCreate public network --------------------------------------------------- 5.94s\nCreate public subnet ---------------------------------------------------- 4.02s\nCreate default IPv4 subnet pool ----------------------------------------- 3.60s\nCreate manager role ----------------------------------------------------- 3.38s\nGathering Facts --------------------------------------------------------- 1.81s\n\u001b[32m2026-01-13 01:13:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 53360f45-5587-4812-b53d-d9a1608bcdef (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m\n\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-01-13 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-01-13 01:13:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m\n\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m\n\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m\n\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-01-13 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-01-13 01:13:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m\n\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m\n\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m\n+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2.CHECKSUM\u001b[0m\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 97638f6d53f99618f0ea9da8ee2e38e3567970c3d42f572e458295ca1456bb27\u001b[0m\n\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87998ab9-56ee-40f7-9801-b62efada077a (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m\n\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2: 200\u001b[0m\n\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m\n\u001b[32m2026-01-13 01:14:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-01-13 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:14:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:14:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:15:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-01-12' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m\n\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m\n\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m\n\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m\n\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m\n\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m\n\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m\n\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m\n\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-01-12\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m\n\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m",
                            "stdout_lines": [
                                "",
                                "# BOOTSTRAP",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# BOOTSTRAP'",
                                "+ echo",
                                "+ sh -c /opt/configuration/scripts/bootstrap-services.sh",
                                "+ set -e",
                                "+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh",
                                "\u001b[32m2026-01-13 01:12:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 65a2231a-56ff-4016-a60d-b8b13fdafb9c (flavor-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c177d464-cf3b-4607-b28c-f60cc9756e07 (bootstrap-basic) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c177d464-cf3b-4607-b28c-f60cc9756e07 (bootstrap-basic) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Bootstrap basic OpenStack services] **************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 13 January 2026  01:12:49 +0000 (0:00:00.065)       0:00:00.065 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Get volume type LUKS] ****************************************************",
                                "Tuesday 13 January 2026  01:12:51 +0000 (0:00:01.806)       0:00:01.872 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Create volume type LUKS] *************************************************",
                                "Tuesday 13 January 2026  01:12:59 +0000 (0:00:08.537)       0:00:10.410 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public network] ***************************************************",
                                "Tuesday 13 January 2026  01:13:07 +0000 (0:00:07.991)       0:00:18.402 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Set public network to default] *******************************************",
                                "Tuesday 13 January 2026  01:13:13 +0000 (0:00:05.944)       0:00:24.346 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public subnet] ****************************************************",
                                "Tuesday 13 January 2026  01:13:19 +0000 (0:00:06.144)       0:00:30.490 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create default IPv4 subnet pool] *****************************************",
                                "Tuesday 13 January 2026  01:13:23 +0000 (0:00:04.015)       0:00:34.506 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create manager role] *****************************************************",
                                "Tuesday 13 January 2026  01:13:27 +0000 (0:00:03.598)       0:00:38.105 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:13:30 +0000 (0:00:03.376)       0:00:41.482 ******* ",
                                "=============================================================================== ",
                                "Get volume type LUKS ---------------------------------------------------- 8.54s",
                                "Create volume type LUKS ------------------------------------------------- 7.99s",
                                "Set public network to default ------------------------------------------- 6.14s",
                                "Create public network --------------------------------------------------- 5.94s",
                                "Create public subnet ---------------------------------------------------- 4.02s",
                                "Create default IPv4 subnet pool ----------------------------------------- 3.60s",
                                "Create manager role ----------------------------------------------------- 3.38s",
                                "Gathering Facts --------------------------------------------------------- 1.81s",
                                "\u001b[32m2026-01-13 01:13:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 53360f45-5587-4812-b53d-d9a1608bcdef (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-01-13 01:13:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2.CHECKSUM\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 97638f6d53f99618f0ea9da8ee2e38e3567970c3d42f572e458295ca1456bb27\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87998ab9-56ee-40f7-9801-b62efada077a (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2: 200\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-01-12' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260112.qcow2\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-01-12\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-01-12'\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-01-13 01:15:26\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001d-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:15:27.289570Z",
                            "start": "2026-01-13T01:12:20.685200Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001d",
                        "name": "Bootstrap services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "/opt/configuration/scripts/check.sh"
                            ],
                            "delta": "0:21:26.306938",
                            "end": "2026-01-13 01:36:53.922637",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 /opt/configuration/scripts/check.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001e-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 01:15:27.615699",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n\n# CHECK\n\n+ echo '# CHECK'\n+ echo\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-manager\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-manager'\n+ echo\n+ osism container testbed-manager ps\n\u001b[32m2026-01-13 01:15:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m\nCONTAINER ID   IMAGE                                                           COMMAND                  CREATED             STATUS                       PORTS                                                                             NAMES\nbeaafb719ec6   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_blackbox_exporter\n1b4b453e1364   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_alertmanager\nd55060940741   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_cadvisor\n76f1ee0cf662   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_node_exporter\n4d3759f39f85   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_server\n507f77bc0f9b   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   17 minutes ago      Up 16 minutes                                                                                                  cephclient\nae1d9bd5208e   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes                                                                                                  cron\n607b2adb2324   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes                                                                                                  kolla_toolbox\nb2a22f1a1da2   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes                                                                                                  fluentd\n9521b89830b6   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   30 minutes ago      Up 30 minutes (healthy)      80/tcp                                                                            phpmyadmin\na6fdd77eefc8   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   31 minutes ago      Up 30 minutes                                                                                                  openstackclient\nd7d4e2e86381   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   31 minutes ago      Up 30 minutes (healthy)      8080/tcp                                                                          homer\n76f921fdade5   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   54 minutes ago      Up 54 minutes (healthy)      192.168.16.5:3128->3128/tcp                                                       squid\ndbd6bf8f01ad   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   58 minutes ago      Up 37 minutes (healthy)                                                                                        manager-inventory_reconciler-1\nb689e6460e5a   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osism-ansible\nea10c30f1a66   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osism-kubernetes\n1e7366c9a3e3   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        ceph-ansible\ne0b054395a95   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        kolla-ansible\na443ed407cc2   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   58 minutes ago      Up 38 minutes (healthy)      8000/tcp                                                                          manager-ara-server-1\n2e264c9269d9   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes                192.168.16.5:3000->3000/tcp                                                       osism-frontend\n3633f57a7562   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)      192.168.16.5:8000->8000/tcp                                                       manager-api-1\neae986e0f082   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osismclient\n6220425b5e3d   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-beat-1\n9bd9a5a8fc79   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes (healthy)      6379/tcp                                                                          manager-redis-1\n28b9850cbc51   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-openstack-1\n67becce7ffcc   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-listener-1\n11f949886b70   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes (healthy)      3306/tcp                                                                          manager-mariadb-1\nb59d339ad122   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-flower-1\nae4ec133e15a   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   About an hour ago   Up About an hour (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik\n\n## Images @ testbed-manager\n\n+ echo\n+ echo '## Images @ testbed-manager'\n+ echo\n+ osism container testbed-manager images\nREPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE\nregistry.osism.tech/osism/osism-ansible                  latest           c94902cef8b2   About an hour ago   611MB\nregistry.osism.tech/osism/kolla-ansible                  2024.2           035febbec78c   About an hour ago   608MB\nregistry.osism.tech/osism/ceph-ansible                   reef             5821b8692edb   About an hour ago   560MB\nregistry.osism.tech/osism/osism                          latest           e6bbfeb96ae9   About an hour ago   384MB\nregistry.osism.tech/osism/osism-kubernetes               latest           ddf1aa7937f6   About an hour ago   1.23GB\nregistry.osism.tech/osism/osism-frontend                 latest           4df35c5f873f   About an hour ago   239MB\nregistry.osism.tech/osism/inventory-reconciler           latest           2cef1bd78903   About an hour ago   335MB\nregistry.osism.tech/osism/openstackclient                2024.2           1f7c62886d13   17 hours ago        238MB\nregistry.osism.tech/osism/cephclient                     reef             762d7b884358   22 hours ago        454MB\nregistry.osism.tech/kolla/cron                           2024.2           18397a23bf8e   23 hours ago        271MB\nregistry.osism.tech/kolla/kolla-toolbox                  2024.2           41bede1e920d   23 hours ago        675MB\nregistry.osism.tech/kolla/fluentd                        2024.2           7623cc118b47   23 hours ago        585MB\nregistry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           61c1e581ebcf   23 hours ago        313MB\nregistry.osism.tech/kolla/prometheus-node-exporter       2024.2           a99e23a01ef0   23 hours ago        311MB\nregistry.osism.tech/kolla/prometheus-v2-server           2024.2           6acc9df66227   23 hours ago        844MB\nregistry.osism.tech/kolla/prometheus-alertmanager        2024.2           9b46a2a48036   23 hours ago        409MB\nregistry.osism.tech/kolla/prometheus-cadvisor            2024.2           aac9a35ae18a   23 hours ago        363MB\nregistry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   5 weeks ago         11.5MB\nregistry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   2 months ago        334MB\nregistry.osism.tech/dockerhub/library/redis              7.4.7-alpine     13105d2858de   2 months ago        41.4MB\nphpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   3 months ago        742MB\nregistry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   4 months ago        275MB\nregistry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   5 months ago        226MB\nregistry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   19 months ago       146MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-0\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\nea188795c6c1   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\n2380e2574883   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n3ae4641f6e4e   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n070dbd4cc6dd   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\n94284355a770   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\n8d4471753615   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\n462cbafa5532   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n7c54a847b027   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api\nd48ceb80e386   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\nd6305862ef7c   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana\n62f14fda8460   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\n239f51dbfcd3   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\ned57ad60b5e1   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\na0c26107ea02   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api\nc62c68d72bdd   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\n04766d8c4ed4   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\n220256b5ba6e   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor\n657a8dd50a44   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter\ndc26c4b469d8   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter\n741d2a945ecf   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter\nd034b962119d   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor\n61a3551be619   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api\nb45c0ccd8ffa   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n8ec28903b5ee   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker\ndb903281ba06   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\nba58d1f35e07   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\ndf783840ea8c   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central\nff822d10bde2   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api\n4af526a74eb1   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api\n9028ed0703eb   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\nd0f93aa70de9   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-0\nb011bf913b67   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker\nb84cfe27a80b   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener\n1b82559e5a07   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api\nb17f16b43353   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\n74106a3d1992   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\n5bc9165738c3   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\ne65e4adc51c6   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             horizon\n622071ae0e63   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb\ne58ff6e6c1bf   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards\na3d338fc1e63   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-0\n77dcf89d1c62   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch\nd06c84aed363   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\ne3a535de601e   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\n98cb6254ccb9   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\nb0f455a2a0fd   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd\nf056a300284a   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_sb_db\n03ee67e9ef7c   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_nb_db\n000f2103ee99   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-0\n5583095e6fa6   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\n8b7e7c971ce6   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             rabbitmq\ne3997df8706a   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n1f42c21a1465   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n25af468da4ed   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db\n0c9311ab9a35   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\ncb2d99d448b9   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\n39b51e8cb6bd   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\nbc1c26491d65   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox\n35ade5107c7a   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-0\n\n+ echo\n+ echo '## Images @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB\nregistry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB\nregistry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB\nregistry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB\nregistry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB\nregistry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB\nregistry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB\nregistry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB\nregistry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB\nregistry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB\nregistry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB\nregistry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB\nregistry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB\nregistry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB\nregistry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB\nregistry.osism.tech/kolla/ceilometer-notification             2024.2    56e7f91c3fab   23 hours ago   981MB\nregistry.osism.tech/kolla/ceilometer-central                  2024.2    e2d1fc7af8d7   23 hours ago   982MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB\nregistry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB\nregistry.osism.tech/kolla/skyline-apiserver                   2024.2    1245eead9e1b   23 hours ago   995MB\nregistry.osism.tech/kolla/skyline-console                     2024.2    02f0b8255576   23 hours ago   1.05GB\nregistry.osism.tech/kolla/aodh-listener                       2024.2    a43a2e0ff586   23 hours ago   980MB\nregistry.osism.tech/kolla/aodh-notifier                       2024.2    464d439ccf30   23 hours ago   980MB\nregistry.osism.tech/kolla/aodh-evaluator                      2024.2    60e1bd484c4b   23 hours ago   980MB\nregistry.osism.tech/kolla/aodh-api                            2024.2    e620ea9760b8   23 hours ago   979MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-1\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-1'\n\n+ echo\n+ osism container testbed-node-1 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n6d070a01e0ba   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\nef7abf888115   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\ne19ed0666c5a   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n1f78bbd2edd7   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\nceaf14fb5ff3   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 4 minutes (healthy)              octavia_api\n499ed120bece   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\nb25e7cdd2871   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n7a0dc4bf0d1a   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana\n355e4bd73e99   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api\na36617691ec6   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n2174ff3d2519   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\n5483ab849afc   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\n6743e3764bd4   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api\ne727d5097483   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\n7fd2e5b353eb   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\n58b0950ef285   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\nbaa8dff29690   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor\n6b1fd66c4011   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter\n336eb7c80b7d   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter\nce571bae73c1   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter\n8b13b6c200fb   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor\n4852346278aa   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api\nb9ae4948f091   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n4bfcaff86703   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker\n9aed24fa714b   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\n1195733f1647   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\n00a03e326baf   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central\n99bf66434f3f   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api\n0957003d948d   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api\n438609e65c4d   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-1\n340d4bad038e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\nb674b86a5c23   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker\n356e680e67cf   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener\n2ac9b005a326   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api\ne2e44ea2a969   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\n03e8c3e57406   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon\nf18cc30a0787   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\n8ba0620e1cfb   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\nc38a7f6e5bd3   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards\n08e025b80e40   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb\nfc97a97867f7   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch\n0fae9b7a4be8   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-1\naea789e7bffb   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\ndc3e7dc2688f   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\nb1897adcd8c4   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\n1b52b52981fa   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd\nf6d9142811d9   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_sb_db\n1d05f2e7c1ae   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_nb_db\n28fc385ae1c1   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\n84ac1be37239   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-1\nb7e3fe639cae   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq\nd3b0f0da827d   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   29 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n04ac4ac09c22   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n67fec7434b5d   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db\n9bd790c0f9a2   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\n243c917244b1   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\n0c552c4f7a35   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\naddc861f1229   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox\n94d921799108   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-1\n\n+ echo\n+ echo '## Images @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB\nregistry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB\nregistry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB\nregistry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB\nregistry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB\nregistry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB\nregistry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB\nregistry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB\nregistry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB\nregistry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB\nregistry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB\nregistry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB\nregistry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB\nregistry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB\nregistry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB\nregistry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n\n## Containers @ testbed-node-2\n\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\na81c4b68fe98   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\nc86f5574a257   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n821a9d333f94   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager\n00bd8e2743a0   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent\nefb72eae51a6   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\nfff5f09e2d33   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\n75b6386ec898   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n170195c7e378   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana\n5ce179670754   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api\ne9fa2a8c6561   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n5a0070b70b0e   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup\need660f315c7   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume\nd4ebfecf6db8   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api\ne0d2df62cfc0   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\neadbd876d04a   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\n642cbc96058d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter\nc007ac101a92   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor\nca7cef65ce90   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter\n759c3a94e90c   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter\n42f151ffaa33   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter\n4f47f34f6acb   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor\n9819d4c0621d   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api\n9ba1c3cfc4e0   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server\n81bb6c51b709   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker\na7cfb21e0460   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns\n677a1f3fdd63   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer\nbb326272f15a   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central\n3dfa91430cf3   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api\n42c3971145b0   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api\nf678a7188e7c   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-2\n420acd86d14e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\na9d4a3c52d07   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker\n29cbcaa8bf53   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener\n94b800366ebb   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api\naae1d7a3322a   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\n454a7bc660a9   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon\nd5c0029d8b42   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\nd0e831db5c33   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\n99beef20bac9   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   21 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards\n3010d9f01137   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb\nc8aa6760e44c   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch\nb5b39c44de9e   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-2\ne5ddb7f2fe7f   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\n8c345e3d480f   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\n6ab1ede542d6   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\n01fc0f6f50b2   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_northd\nec87d6288525   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_sb_db\n54acb50e9e5d   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_nb_db\nfb40cfc2dc34   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq\na69ff4dd3c75   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\n01f2672e35a4   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-2\n07229e97d6f6   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   29 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n5fb37a845f67   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n655e2208ef2e   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db\n517d5e35db12   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\n2e9774371971   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\nc3e66589d1b9   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\n39a23c099af8   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox\nb0f8a3512151   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-2\n\n+ echo\n+ echo '## Images @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB\nregistry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB\nregistry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB\nregistry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB\nregistry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB\nregistry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB\nregistry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB\nregistry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB\nregistry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB\nregistry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB\nregistry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB\nregistry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB\nregistry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB\nregistry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB\nregistry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB\nregistry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB\n+ sh -c /opt/configuration/scripts/check-services.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# Ceph status\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# Ceph status'\n+ echo\n+ ceph -s\n  cluster:\n    id:     11111111-1111-1111-1111-111111111111\n    health: HEALTH_OK\n \n  services:\n    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 27m)\n    mgr: testbed-node-2(active, since 14m), standbys: testbed-node-1, testbed-node-0\n    mds: 1/1 daemons up, 2 standby\n    osd: 6 osds: 6 up (since 23m), 6 in (since 24m)\n    rgw: 3 daemons active (3 hosts, 1 zones)\n \n  data:\n    volumes: 1/1 healthy\n    pools:   14 pools, 401 pgs\n    objects: 552 objects, 2.2 GiB\n    usage:   7.1 GiB used, 113 GiB / 120 GiB avail\n    pgs:     401 active+clean\n \n\n# Ceph versions\n\n+ echo\n+ echo '# Ceph versions'\n+ echo\n+ ceph versions\n{\n    \"mon\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"mgr\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"osd\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6\n    },\n    \"mds\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"rgw\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"overall\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18\n    }\n}\n\n# Ceph OSD tree\n\n+ echo\n+ echo '# Ceph OSD tree'\n+ echo\n+ ceph osd df tree\nID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              \n-1         0.11691         -  120 GiB  7.1 GiB  6.7 GiB    6 KiB  430 MiB  113 GiB  5.92  1.00    -          root default           \n-3         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-3\n 0    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.5 GiB    1 KiB   70 MiB   18 GiB  7.70  1.30  200      up          osd.0          \n 4    hdd  0.01949   1.00000   20 GiB  844 MiB  771 MiB    1 KiB   74 MiB   19 GiB  4.13  0.70  190      up          osd.4          \n-5         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4\n 2    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.4 GiB    1 KiB   74 MiB   19 GiB  7.38  1.25  206      up          osd.2          \n 5    hdd  0.01949   1.00000   20 GiB  912 MiB  843 MiB    1 KiB   70 MiB   19 GiB  4.46  0.75  186      up          osd.5          \n-7         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-5\n 1    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.3 GiB    1 KiB   74 MiB   19 GiB  7.10  1.20  184      up          osd.1          \n 3    hdd  0.01949   1.00000   20 GiB  969 MiB  899 MiB    1 KiB   70 MiB   19 GiB  4.74  0.80  204      up          osd.3          \n                       TOTAL  120 GiB  7.1 GiB  6.7 GiB  9.3 KiB  430 MiB  113 GiB  5.92                                            \nMIN/MAX VAR: 0.70/1.30  STDDEV: 1.50\n\n# Ceph monitor status\n\n+ echo\n+ echo '# Ceph monitor status'\n+ echo\n+ ceph mon stat\ne1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2\n\n# Ceph quorum status\n\n+ echo\n+ echo '# Ceph quorum status'\n+ echo\n+ ceph quorum_status\n+ jq\n{\n  \"election_epoch\": 8,\n  \"quorum\": [\n    0,\n    1,\n    2\n  ],\n  \"quorum_names\": [\n    \"testbed-node-0\",\n    \"testbed-node-1\",\n    \"testbed-node-2\"\n  ],\n  \"quorum_leader_name\": \"testbed-node-0\",\n  \"quorum_age\": 1656,\n  \"features\": {\n    \"quorum_con\": \"4540138322906710015\",\n    \"quorum_mon\": [\n      \"kraken\",\n      \"luminous\",\n      \"mimic\",\n      \"osdmap-prune\",\n      \"nautilus\",\n      \"octopus\",\n      \"pacific\",\n      \"elector-pinging\",\n      \"quincy\",\n      \"reef\"\n    ]\n  },\n  \"monmap\": {\n    \"epoch\": 1,\n    \"fsid\": \"11111111-1111-1111-1111-111111111111\",\n    \"modified\": \"2026-01-13T00:47:57.735690Z\",\n    \"created\": \"2026-01-13T00:47:57.735690Z\",\n    \"min_mon_release\": 18,\n    \"min_mon_release_name\": \"reef\",\n    \"election_strategy\": 1,\n    \"disallowed_leaders: \": \"\",\n    \"stretch_mode\": false,\n    \"tiebreaker_mon\": \"\",\n    \"removed_ranks: \": \"\",\n    \"features\": {\n      \"persistent\": [\n        \"kraken\",\n        \"luminous\",\n        \"mimic\",\n        \"osdmap-prune\",\n        \"nautilus\",\n        \"octopus\",\n        \"pacific\",\n        \"elector-pinging\",\n        \"quincy\",\n        \"reef\"\n      ],\n      \"optional\": []\n    },\n    \"mons\": [\n      {\n        \"rank\": 0,\n        \"name\": \"testbed-node-0\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.10:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.10:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.10:6789/0\",\n        \"public_addr\": \"192.168.16.10:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 1,\n        \"name\": \"testbed-node-1\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.11:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.11:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.11:6789/0\",\n        \"public_addr\": \"192.168.16.11:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 2,\n        \"name\": \"testbed-node-2\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.12:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.12:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.12:6789/0\",\n        \"public_addr\": \"192.168.16.12:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      }\n    ]\n  }\n}\n\n# Ceph free space status\n\n+ echo\n+ echo '# Ceph free space status'\n+ echo\n+ ceph df\n--- RAW STORAGE ---\nCLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED\nhdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92\nTOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92\n \n--- POOLS ---\nPOOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL\n.mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB\ncephfs_data                 2   32      0 B        0      0 B      0     35 GiB\ncephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB\ndefault.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.control         6   32      0 B        8      0 B      0     35 GiB\ndefault.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB\ndefault.rgw.meta            8   32      0 B        0      0 B      0     35 GiB\n.rgw.root                   9   32  1.4 KiB        4   32 KiB      0     52 GiB\nbackups                    10   32     19 B        2   12 KiB      0     35 GiB\nvolumes                    11   32     19 B        2   12 KiB      0     35 GiB\nimages                     12   32  2.2 GiB      299  6.7 GiB   5.98     35 GiB\nmetrics                    13   32     19 B        2   12 KiB      0     35 GiB\nvms                        14   32     19 B        2   12 KiB      0     35 GiB\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ ! -e /etc/redhat-release ]]\n+ osism apply facts\n\u001b[32m2026-01-13 01:16:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a66d0131-6367-4bd5-a2fd-6d1bec74496b (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 01:16:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a66d0131-6367-4bd5-a2fd-6d1bec74496b (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 13 January 2026  01:16:11 +0000 (0:00:00.266)       0:00:00.266 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 13 January 2026  01:16:12 +0000 (0:00:01.377)       0:00:01.644 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 13 January 2026  01:16:13 +0000 (0:00:01.415)       0:00:03.059 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 13 January 2026  01:16:21 +0000 (0:00:07.560)       0:00:10.620 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:16:22 +0000 (0:00:00.545)       0:00:11.166 ******* \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 7.56s\nosism.commons.facts : Copy fact files ----------------------------------- 1.42s\nosism.commons.facts : Create custom facts directory --------------------- 1.38s\nGather facts for all hosts ---------------------------------------------- 0.55s\n+ osism validate ceph-mons\n\nPLAY [Ceph validate mons] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 13 January 2026  01:16:39 +0000 (0:00:00.421)       0:00:00.421 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 13 January 2026  01:16:40 +0000 (0:00:00.777)       0:00:01.198 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 13 January 2026  01:16:41 +0000 (0:00:00.933)       0:00:02.131 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nTuesday 13 January 2026  01:16:41 +0000 (0:00:00.129)       0:00:02.261 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nTuesday 13 January 2026  01:16:41 +0000 (0:00:00.279)       0:00:02.540 ******* \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nTuesday 13 January 2026  01:16:42 +0000 (0:00:00.975)       0:00:03.515 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nTuesday 13 January 2026  01:16:43 +0000 (0:00:00.288)       0:00:03.804 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:16:43 +0000 (0:00:00.485)       0:00:04.289 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mon is not running] ********************\nTuesday 13 January 2026  01:16:44 +0000 (0:00:00.315)       0:00:04.605 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mon is running] ************************\nTuesday 13 January 2026  01:16:44 +0000 (0:00:00.286)       0:00:04.891 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:16:44 +0000 (0:00:00.511)       0:00:05.403 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.263)       0:00:05.667 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.245)       0:00:05.912 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.234)       0:00:06.147 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.068)       0:00:06.215 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.068)       0:00:06.284 ******* \n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:16:45 +0000 (0:00:00.074)       0:00:06.359 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nTuesday 13 January 2026  01:16:46 +0000 (0:00:00.232)       0:00:06.592 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Prepare quorum test vars] ************************************************\nTuesday 13 January 2026  01:16:46 +0000 (0:00:00.257)       0:00:06.850 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get monmap info from one mon container] **********************************\nTuesday 13 January 2026  01:16:46 +0000 (0:00:00.133)       0:00:06.983 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set quorum test data] ****************************************************\nTuesday 13 January 2026  01:16:48 +0000 (0:00:01.696)       0:00:08.680 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail quorum test if not all monitors are in quorum] **********************\nTuesday 13 January 2026  01:16:48 +0000 (0:00:00.519)       0:00:09.200 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass quorum test if all monitors are in quorum] **************************\nTuesday 13 January 2026  01:16:48 +0000 (0:00:00.133)       0:00:09.333 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set fsid test vars] ******************************************************\nTuesday 13 January 2026  01:16:49 +0000 (0:00:00.324)       0:00:09.658 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail Cluster FSID test if FSID does not match configuration] *************\nTuesday 13 January 2026  01:16:49 +0000 (0:00:00.300)       0:00:09.958 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass Cluster FSID test if it matches configuration] **********************\nTuesday 13 January 2026  01:16:49 +0000 (0:00:00.108)       0:00:10.067 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare status test vars] ************************************************\nTuesday 13 January 2026  01:16:49 +0000 (0:00:00.126)       0:00:10.193 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather status data] ******************************************************\nTuesday 13 January 2026  01:16:49 +0000 (0:00:00.108)       0:00:10.301 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set health test data] ****************************************************\nTuesday 13 January 2026  01:16:50 +0000 (0:00:01.243)       0:00:11.544 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable] *************************\nTuesday 13 January 2026  01:16:51 +0000 (0:00:00.298)       0:00:11.843 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if health is acceptable] *****************************\nTuesday 13 January 2026  01:16:51 +0000 (0:00:00.153)       0:00:11.997 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable (strict)] ****************\nTuesday 13 January 2026  01:16:51 +0000 (0:00:00.145)       0:00:12.143 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if status is OK (strict)] ****************************\nTuesday 13 January 2026  01:16:51 +0000 (0:00:00.326)       0:00:12.470 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 13 January 2026  01:16:52 +0000 (0:00:00.149)       0:00:12.619 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 13 January 2026  01:16:52 +0000 (0:00:00.239)       0:00:12.859 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:16:52 +0000 (0:00:00.234)       0:00:13.093 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:16:54 +0000 (0:00:01.750)       0:00:14.843 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:16:54 +0000 (0:00:00.268)       0:00:15.112 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:54 +0000 (0:00:00.258)       0:00:15.370 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:54 +0000 (0:00:00.069)       0:00:15.439 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:16:54 +0000 (0:00:00.084)       0:00:15.523 ******* \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 13 January 2026  01:16:55 +0000 (0:00:00.093)       0:00:15.617 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:16:56 +0000 (0:00:01.482)       0:00:17.099 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-01-13T01:16:40+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:16:57 +0000 (0:00:00.827)       0:00:17.926 ******* \n=============================================================================== \nAggregate test results step one ----------------------------------------- 1.75s\nGet monmap info from one mon container ---------------------------------- 1.70s\nWrite report file ------------------------------------------------------- 1.48s\nGather status data ------------------------------------------------------ 1.24s\nGet container info ------------------------------------------------------ 0.98s\nCreate report output directory ------------------------------------------ 0.93s\nPrint report file information ------------------------------------------- 0.83s\nGet timestamp for report file ------------------------------------------- 0.78s\nSet quorum test data ---------------------------------------------------- 0.52s\nSet test result to passed if ceph-mon is running ------------------------ 0.51s\nSet test result to passed if container is existing ---------------------- 0.49s\nFail cluster-health if health is not acceptable (strict) ---------------- 0.33s\nPass quorum test if all monitors are in quorum -------------------------- 0.32s\nPrepare test data ------------------------------------------------------- 0.32s\nSet fsid test vars ------------------------------------------------------ 0.30s\nSet health test data ---------------------------------------------------- 0.30s\nSet test result to failed if container is missing ----------------------- 0.29s\nSet test result to failed if ceph-mon is not running -------------------- 0.29s\nPrepare test data for container existance test -------------------------- 0.28s\nAggregate test results step two ----------------------------------------- 0.27s\n+ osism validate ceph-mgrs\n\nPLAY [Ceph validate mgrs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 13 January 2026  01:17:09 +0000 (0:00:00.432)       0:00:00.432 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 13 January 2026  01:17:10 +0000 (0:00:00.792)       0:00:01.225 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 13 January 2026  01:17:11 +0000 (0:00:01.036)       0:00:02.261 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nTuesday 13 January 2026  01:17:11 +0000 (0:00:00.130)       0:00:02.392 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nTuesday 13 January 2026  01:17:11 +0000 (0:00:00.286)       0:00:02.678 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nTuesday 13 January 2026  01:17:13 +0000 (0:00:01.251)       0:00:03.930 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nTuesday 13 January 2026  01:17:13 +0000 (0:00:00.273)       0:00:04.203 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:13 +0000 (0:00:00.471)       0:00:04.675 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mgr is not running] ********************\nTuesday 13 January 2026  01:17:14 +0000 (0:00:00.295)       0:00:04.970 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mgr is running] ************************\nTuesday 13 January 2026  01:17:14 +0000 (0:00:00.282)       0:00:05.252 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:17:14 +0000 (0:00:00.494)       0:00:05.747 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.238)       0:00:05.985 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.284)       0:00:06.270 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.251)       0:00:06.521 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.071)       0:00:06.593 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.070)       0:00:06.663 ******* \n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:17:15 +0000 (0:00:00.076)       0:00:06.740 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nTuesday 13 January 2026  01:17:16 +0000 (0:00:00.245)       0:00:06.986 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Define mgr module test vars] *********************************************\nTuesday 13 January 2026  01:17:16 +0000 (0:00:00.253)       0:00:07.239 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather list of mgr modules] **********************************************\nTuesday 13 January 2026  01:17:16 +0000 (0:00:00.115)       0:00:07.355 ******* \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Parse mgr module list from json] *****************************************\nTuesday 13 January 2026  01:17:18 +0000 (0:00:01.834)       0:00:09.189 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Extract list of enabled mgr modules] *************************************\nTuesday 13 January 2026  01:17:18 +0000 (0:00:00.427)       0:00:09.616 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail test if mgr modules are disabled that should be enabled] ************\nTuesday 13 January 2026  01:17:19 +0000 (0:00:00.327)       0:00:09.944 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass test if required mgr modules are enabled] ***************************\nTuesday 13 January 2026  01:17:19 +0000 (0:00:00.124)       0:00:10.069 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 13 January 2026  01:17:19 +0000 (0:00:00.154)       0:00:10.223 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 13 January 2026  01:17:19 +0000 (0:00:00.249)       0:00:10.472 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:17:19 +0000 (0:00:00.232)       0:00:10.704 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:17:20 +0000 (0:00:01.211)       0:00:11.915 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:17:21 +0000 (0:00:00.267)       0:00:12.182 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:21 +0000 (0:00:00.239)       0:00:12.423 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:21 +0000 (0:00:00.068)       0:00:12.491 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:21 +0000 (0:00:00.067)       0:00:12.559 ******* \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 13 January 2026  01:17:21 +0000 (0:00:00.255)       0:00:12.815 ******* \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:17:23 +0000 (0:00:01.299)       0:00:14.115 ******* \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-01-13T01:17:10+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:17:23 +0000 (0:00:00.403)       0:00:14.518 ******* \n=============================================================================== \nGather list of mgr modules ---------------------------------------------- 1.83s\nWrite report file ------------------------------------------------------- 1.30s\nGet container info ------------------------------------------------------ 1.25s\nAggregate test results step one ----------------------------------------- 1.21s\nCreate report output directory ------------------------------------------ 1.04s\nGet timestamp for report file ------------------------------------------- 0.79s\nSet test result to passed if ceph-mgr is running ------------------------ 0.49s\nSet test result to passed if container is existing ---------------------- 0.47s\nParse mgr module list from json ----------------------------------------- 0.43s\nPrint report file information ------------------------------------------- 0.40s\nFlush handlers ---------------------------------------------------------- 0.39s\nExtract list of enabled mgr modules ------------------------------------- 0.33s\nPrepare test data ------------------------------------------------------- 0.30s\nPrepare test data for container existance test -------------------------- 0.29s\nAggregate test results step two ----------------------------------------- 0.28s\nSet test result to failed if ceph-mgr is not running -------------------- 0.28s\nSet test result to failed if container is missing ----------------------- 0.27s\nAggregate test results step two ----------------------------------------- 0.27s\nFail due to missing containers ------------------------------------------ 0.25s\nAggregate test results step three --------------------------------------- 0.25s\n+ osism validate ceph-osds\n\nPLAY [Ceph validate OSDs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 13 January 2026  01:17:40 +0000 (0:00:00.408)       0:00:00.408 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 13 January 2026  01:17:41 +0000 (0:00:00.808)       0:00:01.216 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 13 January 2026  01:17:42 +0000 (0:00:00.499)       0:00:01.716 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 13 January 2026  01:17:42 +0000 (0:00:00.669)       0:00:02.386 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nTuesday 13 January 2026  01:17:43 +0000 (0:00:00.117)       0:00:02.504 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nTuesday 13 January 2026  01:17:43 +0000 (0:00:00.130)       0:00:02.634 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nTuesday 13 January 2026  01:17:43 +0000 (0:00:00.301)       0:00:02.936 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nTuesday 13 January 2026  01:17:43 +0000 (0:00:00.142)       0:00:03.078 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate total number of OSDs in cluster] *******************************\nTuesday 13 January 2026  01:17:43 +0000 (0:00:00.312)       0:00:03.391 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:44 +0000 (0:00:00.601)       0:00:03.992 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers on host] *********************************\nTuesday 13 January 2026  01:17:45 +0000 (0:00:00.480)       0:00:04.472 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'd3a78fcb007e9c9cf8beb00c41c41182ff4f7141b070b40787de7ec0e5920240', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '74ca70dc77d6f055dc1977d50acf32ef5a65859a316ff150c7e956e8ce8f113e', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0f057698e43a39bb26c65a70895321e4c94977d93367f4a34044d6235e50a35d', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '501c2d37b004c04e960539846b79750a31df18d4984494bccb3d321f67b12355', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'fc81a65026c293243876f3a037c928c9e8302c714721c298c36dc8d11fef9f2c', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '16f8e6164b65c1962d79a58cc1df74729f2818837581bec4bb9fd32af3d46c0d', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '77c17e344c751f33b4e619cd49d30ffb165185980a41f698ee4acbf212a87236', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ac50de9f26bbe3e9c8ff977156ab573314956294e9d033aabfec97231c5de069', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'dbf500f7855dadb661620cb1f20e7a118a01cee114c3768e9b58a9de098a0951', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'dcc4c3b0e0ce368d05ffcfe9a38a6a63dbd36bfc0220a82a31c783b587b0d12b', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': '9ff012654489585a85c16e6a37432be319b40455fe8df31375d59993386344b0', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': 'add6cf9e8ba3561a450d02e573c16e4ec90b99d2969871fd5d8194a7d854fb6a', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8c034d9af798b7f0d7980314c29c6fa6f1c01c9ad236cf05c53057267ff850ae', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0476ed090495228a22c133fa90cf3661a6ce0a9edad3173504a702edd4a6f2e7', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'c24bd64d85708f35ec665c98850bf1b7a8be7d31dd6eb78d04f07ea294c03ebc', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e6956e22e953a73a77428962cadb3516942f8cbfd1c7f1e8b8bacd3e2e5d8896', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0a389f80a55f73a57fc38c284bd5b4817cdefaac8ea27b5acb2311f16110e7a7', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ff37babf5b9ed1b3a668f1c5b88f9966d8266abd23565e787200476559530c29', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1de249b56223bcefe828864482a846e0ef2aed56db7e4b7fe0b3f301a8e6a6b6', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '203cae7e9471b4926a0ba5cca1e546eab5204e49a478a8d214ce0305c698ac24', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8ee1be56a0b59d70014e7aacbdd8c7b3e88ef9c11ac1a606f8c66be0478d4d24', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6641d8b49495231b8e1b3f96ca8ee6c0dc613562be2f4df89b8f1a5058b21511', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0eea71216deabba3ffac7a2f099e5eaddb2e82c95eeab999562b2cfaf89b061e', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c3c80c73b6070420d84aee4a793e8abf74aec8f5c2ba8b2ac3f84a94c546ed32', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ef7e548b1eb7fea6e0b310c2dec11236a0162d47a8f555e3963a86a4657d1f01', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c77125e3250d11076354289b7976a700b13354225538c4c521c2e1e903bf8e94', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3ab45e01af15fa611a229b1c189e995bce34019640c57b4b2f6a6cf1b16745c5', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'dd0d63e50b80ccb93e2755428dc7a9d03e7fbfd8406794efafa36fe04fd2d261', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '70a450169183dee9c5f573c57a313c5d7dc19d008fd64ce40f51cfacb51d2998', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': 'd7145a18aee01206e1044557c942e7512c26d36b8ea41dd3bd90dfa1375c494a', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0cda99d5c6691e323e9278321577a2ef44fbe4cd1728ce9f9e010ec19aeb23e9', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e33387bfa156e99b01fde1badacac8ccb7c1e86de42dee06ae8f6248329f3c13', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '25010fc9c54f4efff56dd3d7d36bbbb6978a3d3b714a88e824d15b9670cab64e', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '296655f9405254e03e3e41af4b0c690c30cb7444585d0a8594379f81d663231b', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '20aee76fafcd6639019c80ed9e8e7410768973fbe6af6e91304b80ee51c6eb42', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '761948de6bc91d66b89370e7d945883288207c20eb969377c91dae379d7edd50', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b328d61fdf17323e257ebfd30e1bad731e307c21fecc881c0d2b0b7dbf9b598b', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '23ab7758342847ebd9fb62766ebc2485a7160fc866346d1eefa9cd7ee949d04f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '3ab1060c447cc5406468aef271af340002df988de2302a437d5bb4223e809690', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '0665162f52b763b5c76931e9af0ac5ddca90c79858155af311da5dec36ffae49', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '98b346020f62df488d4e839711b116281e3169623da33161f52db7bf96da431f', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '98dbddf37d2c7c907799c21d6de41884478c4765dce765b1cd20068da70f3006', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4245043ad1a72dbc1b0fd2df2eaf5ca0ebdfb611e01db3906f638ed40bd9d3e0', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9e4bc82239fc81ff290bffbcf7402743ef3d5caacee719ffd53a7d64922b653f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '35a21964a68f180972a4493eee94d2297773e4641b4a8b5fc45aa524d89c0331', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1e933cbe1b7199bdb6c7dc8be19f3f068c99fe685ce5ca443ba0040fd5129de7', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '5b124b584b96f6e367fb9719c890f8fe2a7423c3b676041d3b94fe8c9864fac8', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '665fd7911952682cc2e6c6c1adc4e90293e5a58e0dc18cbdc6f89c6689f2ea75', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'dded726ea7cd20156756fac7346c2c27432e38b646e04072a7add8bd5f9bce38', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fea967d64cbb6e2e73642576dc14f4c869b4afcc8691d2d277c84322fe2d1d3a', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9c6af98f19ef3213321fa58afcf4a625c5bcb6557e92f5a59f835066551aa5a6', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '3890e7b35cfca588e566c42f4a3a874bf3e39e84e26e89b2a52a088438fbcfac', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8d0fd15f8d0ca6eda9ff73dab70c2224170befd946471b96881cf334b7b18763', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '14dfc3f860f2cfff4737b8970309e3f5d84a8d27141a4b67b7aeb9c653936983', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\nTASK [Get count of ceph-osd containers on host] ********************************\nTuesday 13 January 2026  01:17:45 +0000 (0:00:00.480)       0:00:04.953 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed when count of containers is wrong] *************\nTuesday 13 January 2026  01:17:45 +0000 (0:00:00.301)       0:00:05.254 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if count matches] ******************************\nTuesday 13 January 2026  01:17:46 +0000 (0:00:00.494)       0:00:05.749 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:46 +0000 (0:00:00.286)       0:00:06.035 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers that are not running] ********************\nTuesday 13 January 2026  01:17:46 +0000 (0:00:00.276)       0:00:06.312 ******* \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get count of ceph-osd containers that are not running] *******************\nTuesday 13 January 2026  01:17:47 +0000 (0:00:00.300)       0:00:06.612 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nTuesday 13 January 2026  01:17:47 +0000 (0:00:00.473)       0:00:07.086 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nTuesday 13 January 2026  01:17:47 +0000 (0:00:00.294)       0:00:07.381 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if all containers are running] *****************\nTuesday 13 January 2026  01:17:48 +0000 (0:00:00.275)       0:00:07.656 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:17:48 +0000 (0:00:00.277)       0:00:07.933 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:17:48 +0000 (0:00:00.470)       0:00:08.404 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:17:49 +0000 (0:00:00.626)       0:00:09.030 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:49 +0000 (0:00:00.231)       0:00:09.261 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:49 +0000 (0:00:00.068)       0:00:09.329 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:17:49 +0000 (0:00:00.080)       0:00:09.410 ******* \n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:17:50 +0000 (0:00:00.083)       0:00:09.493 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail early due to containers not running] ********************************\nTuesday 13 January 2026  01:17:50 +0000 (0:00:00.251)       0:00:09.744 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:50 +0000 (0:00:00.238)       0:00:09.983 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set _mon_hostname fact] **************************************************\nTuesday 13 January 2026  01:17:50 +0000 (0:00:00.290)       0:00:10.274 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get ceph osd tree] *******************************************************\nTuesday 13 January 2026  01:17:51 +0000 (0:00:00.239)       0:00:10.513 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Parse osd tree from JSON] ************************************************\nTuesday 13 January 2026  01:17:53 +0000 (0:00:02.142)       0:00:12.655 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get OSDs that are not up or in] ******************************************\nTuesday 13 January 2026  01:17:53 +0000 (0:00:00.132)       0:00:12.787 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Fail test if OSDs are not up or in] **************************************\nTuesday 13 January 2026  01:17:53 +0000 (0:00:00.289)       0:00:13.077 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Pass test if OSDs are all up and in] *************************************\nTuesday 13 January 2026  01:17:53 +0000 (0:00:00.121)       0:00:13.198 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:53 +0000 (0:00:00.108)       0:00:13.306 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [List ceph LVM volumes and collect data] **********************************\nTuesday 13 January 2026  01:17:54 +0000 (0:00:00.281)       0:00:13.588 ******* \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Parse LVM data as JSON] **************************************************\nTuesday 13 January 2026  01:17:56 +0000 (0:00:02.739)       0:00:16.328 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get unencrypted and encrypted OSDs] **************************************\nTuesday 13 January 2026  01:17:57 +0000 (0:00:00.484)       0:00:16.812 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of encrypted OSDs does not match] **************************\nTuesday 13 January 2026  01:17:57 +0000 (0:00:00.478)       0:00:17.291 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of encrypted OSDs equals count of OSDs] ********************\nTuesday 13 January 2026  01:17:58 +0000 (0:00:00.290)       0:00:17.581 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of unencrypted OSDs does not match] ************************\nTuesday 13 January 2026  01:17:58 +0000 (0:00:00.461)       0:00:18.043 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************\nTuesday 13 January 2026  01:17:58 +0000 (0:00:00.276)       0:00:18.319 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 13 January 2026  01:17:59 +0000 (0:00:00.263)       0:00:18.583 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get CRUSH node data of each OSD host and root node childs] ***************\nTuesday 13 January 2026  01:17:59 +0000 (0:00:00.547)       0:00:19.130 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate sub test expression results] ***********************************\nTuesday 13 January 2026  01:18:00 +0000 (0:00:00.903)       0:00:20.034 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail test if any sub test failed] ****************************************\nTuesday 13 January 2026  01:18:00 +0000 (0:00:00.296)       0:00:20.331 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass test if no sub test failed] *****************************************\nTuesday 13 January 2026  01:18:01 +0000 (0:00:00.268)       0:00:20.599 ******* \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 13 January 2026  01:18:01 +0000 (0:00:00.287)       0:00:20.887 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 13 January 2026  01:18:01 +0000 (0:00:00.234)       0:00:21.122 ******* \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 13 January 2026  01:18:02 +0000 (0:00:00.694)       0:00:21.816 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 13 January 2026  01:18:03 +0000 (0:00:01.512)       0:00:23.329 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 13 January 2026  01:18:04 +0000 (0:00:00.278)       0:00:23.607 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:18:04 +0000 (0:00:00.230)       0:00:23.838 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:18:04 +0000 (0:00:00.068)       0:00:23.906 ******* \n\nTASK [Flush handlers] **********************************************************\nTuesday 13 January 2026  01:18:04 +0000 (0:00:00.066)       0:00:23.972 ******* \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 13 January 2026  01:18:04 +0000 (0:00:00.069)       0:00:24.041 ******* \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 13 January 2026  01:18:05 +0000 (0:00:01.240)       0:00:25.282 ******* \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-01-13T01:17:41+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:18:06 +0000 (0:00:00.386)       0:00:25.668 ******* \n=============================================================================== \nList ceph LVM volumes and collect data ---------------------------------- 2.74s\nGet ceph osd tree ------------------------------------------------------- 2.14s\nAggregate test results step one ----------------------------------------- 1.51s\nWrite report file ------------------------------------------------------- 1.24s\nGet CRUSH node data of each OSD host and root node childs --------------- 0.90s\nGet timestamp for report file ------------------------------------------- 0.81s\nSet validation result to failed if a test failed ------------------------ 0.69s\nCreate report output directory ------------------------------------------ 0.67s\nAggregate test results step two ----------------------------------------- 0.63s\nCalculate total number of OSDs in cluster ------------------------------- 0.60s\nPrepare test data ------------------------------------------------------- 0.55s\nGet extra vars for Ceph configuration ----------------------------------- 0.50s\nSet test result to failed when count of containers is wrong ------------- 0.49s\nParse LVM data as JSON -------------------------------------------------- 0.48s\nPrepare test data ------------------------------------------------------- 0.48s\nGet list of ceph-osd containers on host --------------------------------- 0.48s\nGet unencrypted and encrypted OSDs -------------------------------------- 0.48s\nGet count of ceph-osd containers that are not running ------------------- 0.47s\nAggregate test results step one ----------------------------------------- 0.47s\nPass if count of encrypted OSDs equals count of OSDs -------------------- 0.46s\n+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ -e /etc/redhat-release ]]\n+ source /etc/os-release\n++ PRETTY_NAME='Ubuntu 24.04.3 LTS'\n++ NAME=Ubuntu\n++ VERSION_ID=24.04\n++ VERSION='24.04.3 LTS (Noble Numbat)'\n++ VERSION_CODENAME=noble\n++ ID=ubuntu\n++ ID_LIKE=debian\n++ HOME_URL=https://www.ubuntu.com/\n++ SUPPORT_URL=https://help.ubuntu.com/\n++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/\n++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\n++ UBUNTU_CODENAME=noble\n++ LOGO=ubuntu-logo\n+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]\n+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'\n+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n\n# Status of Elasticsearch\n\n+ pushd /opt/configuration/contrib\n+ echo\n+ echo '# Status of Elasticsearch'\n+ echo\n+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s\nOK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0\n\n+ echo\n+ echo '# Status of MariaDB'\n# Status of MariaDB\n\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status database\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: 4c7e25da-f01a-11f0-8988-c257ed0aca95\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.24(ra6b53429)\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: 81b21e74-f01a-11f0-9b30-e2d7a1171da6\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000127307\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 5073 local commits, 7788 replicated, 122 received\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 23 minutes, 23 seconds\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 128 connected, 1 running\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 137868 total, 0 slow\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 42\u001b[0m\n\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m\n\n# Status of Prometheus\n\n+ echo\n+ echo '# Status of Prometheus'\n+ echo\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy\nUnauthorized\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready\nUnauthorized\n\n# Status of RabbitMQ\n\n+ echo\n+ echo '# Status of RabbitMQ'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status messaging\n\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 204, Channels: 203, Queues: 173\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 119/1024\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 71/832\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 204, Channels: 203, Queues: 173\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.8 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 124/1024\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 78/832\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 204, Channels: 203, Queues: 173\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.9 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 101/1024\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 55/832\u001b[0m\n\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m\n\n# Status of Redis\n\n+ echo\n+ echo '# Status of Redis'\n+ echo\n+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j\nTCP OK - 0.001 second response time on 192.168.16.10 port 6379|time=0.001242s;;;0.000000;10.000000\n+ popd\n\n# Create backup of MariaDB database\n\n+ echo\n+ echo '# Create backup of MariaDB database'\n+ echo\n+ osism apply mariadb_backup -e mariadb_backup_type=full\n\u001b[32m2026-01-13 01:18:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 062ddd45-8a86-4a8a-9d97-879af9c04429 (mariadb_backup) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 01:18:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 062ddd45-8a86-4a8a-9d97-879af9c04429 (mariadb_backup) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 13 January 2026  01:19:00 +0000 (0:00:00.168)       0:00:00.168 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 13 January 2026  01:19:00 +0000 (0:00:00.321)       0:00:00.490 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nTuesday 13 January 2026  01:19:00 +0000 (0:00:00.566)       0:00:01.057 ******* \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 13 January 2026  01:19:01 +0000 (0:00:00.413)       0:00:01.470 ******* \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Get MariaDB container facts] ***********************************\nTuesday 13 January 2026  01:19:01 +0000 (0:00:00.541)       0:00:02.012 ******* \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Taking full database backup via Mariabackup] *******************\nTuesday 13 January 2026  01:19:05 +0000 (0:00:03.276)       0:00:05.289 ******* \n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nTuesday 13 January 2026  01:19:46 +0000 (0:00:41.685)       0:00:46.974 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nTuesday 13 January 2026  01:19:47 +0000 (0:00:00.299)       0:00:47.274 ******* \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:19:47 +0000 (0:00:00.393)       0:00:47.668 ******* \n=============================================================================== \nmariadb : Taking full database backup via Mariabackup ------------------ 41.69s\nmariadb : Get MariaDB container facts ----------------------------------- 3.28s\nGroup hosts based on enabled services ----------------------------------- 0.57s\nmariadb : include_tasks ------------------------------------------------- 0.54s\nmariadb : Group MariaDB hosts based on shards --------------------------- 0.41s\nInclude mariadb post-upgrade.yml ---------------------------------------- 0.39s\nGroup hosts based on Kolla action --------------------------------------- 0.32s\nInclude mariadb post-deploy.yml ----------------------------------------- 0.30s\n+ sh -c /opt/configuration/scripts/check/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# OpenStack endpoints\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ export OS_CLOUD=admin\n+ OS_CLOUD=admin\n+ echo\n+ echo '# OpenStack endpoints'\n+ echo\n+ openstack endpoint list\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| 06574ad6874b4068b9bc0715f33ce4d8 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |\n| 19208f7c81644839bb2114dd1b2aa712 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |\n| 256c014a163d4bb3b0a21d6d415e3f1f | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |\n| 27ffb9623e47436c9a7b873bfe49df78 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |\n| 2938964b65df47a2b60afc351c312186 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |\n| 3556e304fef940c7ac51bba018c10a60 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |\n| 462fe6fc292349acba4e5368c004e13f | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |\n| 582285c0b7354a218609896e63a7c1ca | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |\n| 61be64e294ac46e487d30d3c3dbc5a80 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |\n| 620a43c623624295a911aabcdc22cff3 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |\n| 87a2354dffd945359975254d4d61b7a0 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |\n| 896f8523f4b14bf59f1dae3a360b1340 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |\n| 9d0b4c8fe69f4495be961f98e0af2b7a | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |\n| b437ac5f277d43388ed5f54aca4d5319 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |\n| b4af2bdf24e9452293a735ab95ecb545 | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |\n| bb9a954b184546e4a6cbfc27a68b076a | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |\n| c5df2317bce0469fb96e3c128fa439a3 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |\n| c7487228917044509d34aef0ae5e2e99 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |\n| d5e461332ec449ed9e1a8039c62648d1 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |\n| dab0159c02e54dd38966f555de81adab | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |\n| df150eb0270d4d5c97a7ce8e4aac8bb7 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |\n| f1d61a022aaf45de8479808af7cbc2c0 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n\n# Cinder\n\n+ echo\n+ echo '# Cinder'\n+ echo\n+ openstack volume service list\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| Binary           | Host                       | Zone     | Status  | State | Updated At                 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |\n| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |\n| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |\n| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:49.000000 |\n| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:51.000000 |\n| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:52.000000 |\n| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-01-13T01:19:44.000000 |\n| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-01-13T01:19:45.000000 |\n| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-01-13T01:19:46.000000 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n\n# Neutron\n\n+ echo\n+ echo '# Neutron'\n+ echo\n+ openstack network agent list\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |\n| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n+ openstack network service provider list\n+---------------+------+---------+\n| Service Type  | Name | Default |\n+---------------+------+---------+\n| L3_ROUTER_NAT | ovn  | True    |\n+---------------+------+---------+\n\n# Nova\n\n+ echo\n+ echo '# Nova'\n+ echo\n+ openstack compute service list\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ad6e1ef3-4b1f-47bb-a5e8-f20d1f9715e8 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-01-13T01:19:55.000000 |\n| 02ecf3a3-9236-4c9e-a41f-4511b5e69fae | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-01-13T01:19:56.000000 |\n| e20dd9f4-0b2d-4012-8452-dcbf772284e2 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-01-13T01:19:57.000000 |\n| 6571e646-fa52-46f1-a7ee-4c48984f15e4 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-01-13T01:19:59.000000 |\n| 5aeeec1b-e06e-4ab7-bcd5-3e7efb785f72 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-01-13T01:20:00.000000 |\n| 3f0b0182-a9a1-42f3-b8ed-0f2b28e43402 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-01-13T01:20:00.000000 |\n| 29e50209-78a5-4012-a38a-bca2851525b0 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-01-13T01:20:00.000000 |\n| bb892313-8417-4ff5-8c2f-f9bf9aa0b005 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-01-13T01:20:00.000000 |\n| 3d63b899-2787-417e-96ce-91b3e6f96fad | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-01-13T01:20:01.000000 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n+ openstack hypervisor list\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| 8f93afd9-4cc1-49f6-82ec-3a6527093679 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |\n| cbb720b7-82d5-452e-bf67-987b544a2ead | testbed-node-3      | QEMU            | 192.168.16.13 | up    |\n| 1269b2e9-7f60-43bd-a2a9-9cd7fb028269 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n+ echo\n\n# Run OpenStack test play\n\n+ echo '# Run OpenStack test play'\n+ echo\n+ osism apply --environment openstack test\n\u001b[32m2026-01-13 01:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m\n\u001b[32m2026-01-13 01:20:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0889392d-28b6-4335-8b89-a3e9cb5f251a (test) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 01:20:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0889392d-28b6-4335-8b89-a3e9cb5f251a (test) has been started and output is visible here.\u001b[0m\n\nPLAY [Create test project] *****************************************************\n\nTASK [Create test domain] ******************************************************\nTuesday 13 January 2026  01:20:22 +0000 (0:00:00.051)       0:00:00.051 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test-admin user] **************************************************\nTuesday 13 January 2026  01:20:24 +0000 (0:00:02.834)       0:00:02.885 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add manager role to user test-admin] *************************************\nTuesday 13 January 2026  01:20:28 +0000 (0:00:03.758)       0:00:06.644 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test project] *****************************************************\nTuesday 13 January 2026  01:20:35 +0000 (0:00:06.479)       0:00:13.123 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test user] ********************************************************\nTuesday 13 January 2026  01:20:39 +0000 (0:00:03.976)       0:00:17.100 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add member roles to user test] *******************************************\nTuesday 13 January 2026  01:20:43 +0000 (0:00:03.934)       0:00:21.034 ******* \n\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m\n\nTASK [Create test server group] ************************************************\nTuesday 13 January 2026  01:20:54 +0000 (0:00:11.122)       0:00:32.156 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create ssh security group] ***********************************************\nTuesday 13 January 2026  01:20:58 +0000 (0:00:04.112)       0:00:36.269 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to ssh security group] ******************************************\nTuesday 13 January 2026  01:21:03 +0000 (0:00:04.889)       0:00:41.159 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create icmp security group] **********************************************\nTuesday 13 January 2026  01:21:07 +0000 (0:00:04.071)       0:00:45.231 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to icmp security group] *****************************************\nTuesday 13 January 2026  01:21:11 +0000 (0:00:03.969)       0:00:49.200 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test keypair] *****************************************************\nTuesday 13 January 2026  01:21:15 +0000 (0:00:04.176)       0:00:53.377 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test network] *****************************************************\nTuesday 13 January 2026  01:21:19 +0000 (0:00:03.732)       0:00:57.110 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test subnet] ******************************************************\nTuesday 13 January 2026  01:21:23 +0000 (0:00:04.466)       0:01:01.577 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test router] ******************************************************\nTuesday 13 January 2026  01:21:29 +0000 (0:00:05.592)       0:01:07.169 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test instances] ***************************************************\nTuesday 13 January 2026  01:21:41 +0000 (0:00:11.733)       0:01:18.902 ******* \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\nSTILL ALIVE [task 'Create test instances' is running] **************************\n\nSTILL ALIVE [task 'Create test instances' is running] **************************\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\nSTILL ALIVE [task 'Create test instances' is running] **************************\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\nSTILL ALIVE [task 'Create test instances' is running] **************************\n\nSTILL ALIVE [task 'Create test instances' is running] **************************\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Add metadata to instances] ***********************************************\nTuesday 13 January 2026  01:25:55 +0000 (0:04:14.951)       0:05:33.854 ******* \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Add tag to instances] ****************************************************\nTuesday 13 January 2026  01:26:18 +0000 (0:00:22.113)       0:05:55.968 ******* \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Create test volume] ******************************************************\nTuesday 13 January 2026  01:26:51 +0000 (0:00:33.491)       0:06:29.460 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Attach test volume] ******************************************************\nTuesday 13 January 2026  01:26:58 +0000 (0:00:06.858)       0:06:36.318 ******* \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create floating ip address] **********************************************\nTuesday 13 January 2026  01:27:11 +0000 (0:00:13.010)       0:06:49.328 ******* \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Print floating ip address] ***********************************************\nTuesday 13 January 2026  01:27:17 +0000 (0:00:05.645)       0:06:54.974 ******* \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"192.168.112.133\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:27:17 +0000 (0:00:00.036)       0:06:55.010 ******* \n=============================================================================== \nCreate test instances ------------------------------------------------- 254.95s\nAdd tag to instances --------------------------------------------------- 33.49s\nAdd metadata to instances ---------------------------------------------- 22.11s\nAttach test volume ----------------------------------------------------- 13.01s\nCreate test router ----------------------------------------------------- 11.73s\nAdd member roles to user test ------------------------------------------ 11.12s\nCreate test volume ------------------------------------------------------ 6.86s\nAdd manager role to user test-admin ------------------------------------- 6.48s\nCreate floating ip address ---------------------------------------------- 5.65s\nCreate test subnet ------------------------------------------------------ 5.59s\nCreate ssh security group ----------------------------------------------- 4.89s\nCreate test network ----------------------------------------------------- 4.47s\nAdd rule to icmp security group ----------------------------------------- 4.18s\nCreate test server group ------------------------------------------------ 4.11s\nAdd rule to ssh security group ------------------------------------------ 4.07s\nCreate test project ----------------------------------------------------- 3.98s\nCreate icmp security group ---------------------------------------------- 3.97s\nCreate test user -------------------------------------------------------- 3.93s\nCreate test-admin user -------------------------------------------------- 3.76s\nCreate test keypair ----------------------------------------------------- 3.73s\n+ server_list\n+ openstack --os-cloud test server list\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE | test=192.168.112.182, 192.168.200.225 | N/A (booted from volume) | SCS-1L-1 |\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE | test=192.168.112.189, 192.168.200.82  | N/A (booted from volume) | SCS-1L-1 |\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE | test=192.168.112.197, 192.168.200.207 | N/A (booted from volume) | SCS-1L-1 |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE | test=192.168.112.184, 192.168.200.155 | N/A (booted from volume) | SCS-1L-1 |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE | test=192.168.112.133, 192.168.200.13  | N/A (booted from volume) | SCS-1L-1 |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n+ openstack --os-cloud test server show test\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-01-13T01:22:25.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.133, 192.168.200.13                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-01-13T01:21:49Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | b7c58971654cc5a37db85f5dfcac31fb8cab3c23a991b40125a50d5c                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 105b5d47-9d05-45ab-8759-11299bd19793                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-01-13T01:26:00Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='9a1dd45e-c084-46c9-9b33-74ce3d98308a'                                                                                                                                                                                                                                                                                                             |\n|                                     | delete_on_termination='False', id='6b4757db-f21e-47fa-a113-8d503701f961'                                                                                                                                                                                                                                                                                                            |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-1\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-01-13T01:23:22.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.184, 192.168.200.155                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-01-13T01:22:48Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 299cc453120c89f48d302e3526bfcb16d08ef82d89c546a2b6cd9172                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 6dc66e2a-091f-4945-a07c-d12a01a3a7a2                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-01-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='6f2b9777-0ed8-44e1-ab2e-af2a66da4310'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-2\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-01-13T01:24:18.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.197, 192.168.200.207                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-01-13T01:23:43Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 3b5320a87cce66fed3f5927f5afa231c08cad5c9a8f54dc3b7cadf91                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | e7fd2462-0b34-4225-997f-2488fa25b38e                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-01-13T01:26:08Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='7da2bc0c-af16-4147-95c4-154cd1481cc8'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-3\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-01-13T01:25:01.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.189, 192.168.200.82                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-01-13T01:24:36Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 299cc453120c89f48d302e3526bfcb16d08ef82d89c546a2b6cd9172                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 8ac30647-bb69-4d10-9174-753873ad881b                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-01-13T01:26:13Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='3fa8b44a-0e91-46fa-a4f5-3b822a805fe8'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-4\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-01-13T01:25:43.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.182, 192.168.200.225                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-01-13T01:25:18Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | b7c58971654cc5a37db85f5dfcac31fb8cab3c23a991b40125a50d5c                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 1b15012c-1cbf-4117-a3aa-ee944b91d5d2                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-01-13T01:26:17Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='484c345a-1c17-48f4-8619-af4ffc6b78e7'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.197\nPING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.\n64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=9.85 ms\n64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.53 ms\n64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.71 ms\n\n--- 192.168.112.197 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.714/4.698/9.853/3.660 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=7.24 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=3.01 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.33 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.334/3.863/7.243/2.486 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.66 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.89 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.30 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.298/2.947/5.658/1.931 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.189\nPING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.\n64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=5.29 ms\n64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.75 ms\n64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.53 ms\n\n--- 192.168.112.189 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.525/2.855/5.291/1.724 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.133\nPING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.\n64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.22 ms\n64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.68 ms\n64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.38 ms\n\n--- 192.168.112.133 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.375/2.424/4.216/1.273 ms\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ compute_list\n+ osism manage compute list testbed-node-3\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4\n\u001b[32m2026-01-13 01:28:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m\n\u001b[32m2026-01-13 01:28:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:28:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m\n\u001b[32m2026-01-13 01:28:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:28:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.197\nPING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.\n64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=7.49 ms\n64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=1.61 ms\n64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.34 ms\n\n--- 192.168.112.197 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.341/3.482/7.494/2.838 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=4.36 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=1.88 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.54 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.536/2.592/4.360/1.257 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.10 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.41 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.17 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.171/2.227/4.100/1.327 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.189\nPING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.\n64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=6.53 ms\n64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.64 ms\n64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.39 ms\n\n--- 192.168.112.189 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.391/3.188/6.533/2.367 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.133\nPING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.\n64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.03 ms\n64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.37 ms\n64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.24 ms\n\n--- 192.168.112.133 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.243/2.213/4.026/1.282 ms\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5\n\u001b[32m2026-01-13 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m\n\u001b[32m2026-01-13 01:29:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:29:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m\n\u001b[32m2026-01-13 01:30:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:30:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.197\nPING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.\n64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=5.99 ms\n64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.77 ms\n64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.85 ms\n\n--- 192.168.112.197 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.846/3.535/5.993/1.778 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.14 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=1.60 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.36 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.363/3.032/6.136/2.196 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=3.73 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.89 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.32 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.318/2.315/3.734/1.030 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.189\nPING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.\n64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=4.87 ms\n64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.51 ms\n64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.13 ms\n\n--- 192.168.112.189 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.134/2.505/4.871/1.679 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.133\nPING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.\n64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=3.11 ms\n64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=2.25 ms\n64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.70 ms\n\n--- 192.168.112.133 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.701/2.354/3.111/0.580 ms\n+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3\n\u001b[32m2026-01-13 01:31:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m\n\u001b[32m2026-01-13 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m\n\u001b[32m2026-01-13 01:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:32:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e7fd2462-0b34-4225-997f-2488fa25b38e\u001b[0m\n\u001b[32m2026-01-13 01:32:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:32:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m\n\u001b[32m2026-01-13 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:32:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:33:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m\n\u001b[32m2026-01-13 01:33:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:33:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.197\nPING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.\n64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=4.01 ms\n64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.79 ms\n64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=2.05 ms\n\n--- 192.168.112.197 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.053/2.952/4.012/0.807 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.93 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.04 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.46 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 2.043/3.812/6.933/2.213 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=9.59 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.55 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.24 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.244/4.460/9.590/3.666 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.189\nPING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.\n64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=3.65 ms\n64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.39 ms\n64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.19 ms\n\n--- 192.168.112.189 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.191/2.075/3.646/1.113 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.133\nPING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.\n64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=3.97 ms\n64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.71 ms\n64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.58 ms\n\n--- 192.168.112.133 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.579/2.418/3.965/1.095 ms\n+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4\n\u001b[32m2026-01-13 01:34:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m\n\u001b[32m2026-01-13 01:34:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m\n\u001b[32m2026-01-13 01:34:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:34:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:35:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e7fd2462-0b34-4225-997f-2488fa25b38e\u001b[0m\n\u001b[32m2026-01-13 01:35:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m\n\u001b[32m2026-01-13 01:35:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:35:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-01-13 01:35:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m\n\u001b[32m2026-01-13 01:36:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m\n\u001b[32m2026-01-13 01:36:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |\n| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |\n| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |\n| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |\n| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.197\nPING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.\n64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=8.48 ms\n64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.13 ms\n64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.60 ms\n\n--- 192.168.112.197 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.596/4.068/8.483/3.129 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=4.66 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.29 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.52 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.522/2.825/4.659/1.334 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=3.96 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.01 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.74 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.738/2.567/3.956/0.988 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.189\nPING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.\n64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=4.39 ms\n64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=2.19 ms\n64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.76 ms\n\n--- 192.168.112.189 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.763/2.782/4.391/1.151 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.133\nPING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.\n64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.42 ms\n64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.78 ms\n64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.83 ms\n\n--- 192.168.112.133 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.776/2.674/4.415/1.230 ms",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "",
                                "# CHECK",
                                "",
                                "+ echo '# CHECK'",
                                "+ echo",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-manager",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager ps",
                                "\u001b[32m2026-01-13 01:15:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m",
                                "CONTAINER ID   IMAGE                                                           COMMAND                  CREATED             STATUS                       PORTS                                                                             NAMES",
                                "beaafb719ec6   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_blackbox_exporter",
                                "1b4b453e1364   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_alertmanager",
                                "d55060940741   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   11 minutes ago      Up 11 minutes                                                                                                  prometheus_cadvisor",
                                "76f1ee0cf662   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_node_exporter",
                                "4d3759f39f85   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_server",
                                "507f77bc0f9b   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   17 minutes ago      Up 16 minutes                                                                                                  cephclient",
                                "ae1d9bd5208e   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes                                                                                                  cron",
                                "607b2adb2324   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   29 minutes ago      Up 29 minutes                                                                                                  kolla_toolbox",
                                "b2a22f1a1da2   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   30 minutes ago      Up 30 minutes                                                                                                  fluentd",
                                "9521b89830b6   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   30 minutes ago      Up 30 minutes (healthy)      80/tcp                                                                            phpmyadmin",
                                "a6fdd77eefc8   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   31 minutes ago      Up 30 minutes                                                                                                  openstackclient",
                                "d7d4e2e86381   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   31 minutes ago      Up 30 minutes (healthy)      8080/tcp                                                                          homer",
                                "76f921fdade5   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   54 minutes ago      Up 54 minutes (healthy)      192.168.16.5:3128->3128/tcp                                                       squid",
                                "dbd6bf8f01ad   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   58 minutes ago      Up 37 minutes (healthy)                                                                                        manager-inventory_reconciler-1",
                                "b689e6460e5a   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osism-ansible",
                                "ea10c30f1a66   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osism-kubernetes",
                                "1e7366c9a3e3   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        ceph-ansible",
                                "e0b054395a95   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        kolla-ansible",
                                "a443ed407cc2   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   58 minutes ago      Up 38 minutes (healthy)      8000/tcp                                                                          manager-ara-server-1",
                                "2e264c9269d9   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes                192.168.16.5:3000->3000/tcp                                                       osism-frontend",
                                "3633f57a7562   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)      192.168.16.5:8000->8000/tcp                                                       manager-api-1",
                                "eae986e0f082   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        osismclient",
                                "6220425b5e3d   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-beat-1",
                                "9bd9a5a8fc79   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes (healthy)      6379/tcp                                                                          manager-redis-1",
                                "28b9850cbc51   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-openstack-1",
                                "67becce7ffcc   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-listener-1",
                                "11f949886b70   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   58 minutes ago      Up 38 minutes (healthy)      3306/tcp                                                                          manager-mariadb-1",
                                "b59d339ad122   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago      Up 38 minutes (healthy)                                                                                        manager-flower-1",
                                "ae4ec133e15a   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   About an hour ago   Up About an hour (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik",
                                "",
                                "## Images @ testbed-manager",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager images",
                                "REPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE",
                                "registry.osism.tech/osism/osism-ansible                  latest           c94902cef8b2   About an hour ago   611MB",
                                "registry.osism.tech/osism/kolla-ansible                  2024.2           035febbec78c   About an hour ago   608MB",
                                "registry.osism.tech/osism/ceph-ansible                   reef             5821b8692edb   About an hour ago   560MB",
                                "registry.osism.tech/osism/osism                          latest           e6bbfeb96ae9   About an hour ago   384MB",
                                "registry.osism.tech/osism/osism-kubernetes               latest           ddf1aa7937f6   About an hour ago   1.23GB",
                                "registry.osism.tech/osism/osism-frontend                 latest           4df35c5f873f   About an hour ago   239MB",
                                "registry.osism.tech/osism/inventory-reconciler           latest           2cef1bd78903   About an hour ago   335MB",
                                "registry.osism.tech/osism/openstackclient                2024.2           1f7c62886d13   17 hours ago        238MB",
                                "registry.osism.tech/osism/cephclient                     reef             762d7b884358   22 hours ago        454MB",
                                "registry.osism.tech/kolla/cron                           2024.2           18397a23bf8e   23 hours ago        271MB",
                                "registry.osism.tech/kolla/kolla-toolbox                  2024.2           41bede1e920d   23 hours ago        675MB",
                                "registry.osism.tech/kolla/fluentd                        2024.2           7623cc118b47   23 hours ago        585MB",
                                "registry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           61c1e581ebcf   23 hours ago        313MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter       2024.2           a99e23a01ef0   23 hours ago        311MB",
                                "registry.osism.tech/kolla/prometheus-v2-server           2024.2           6acc9df66227   23 hours ago        844MB",
                                "registry.osism.tech/kolla/prometheus-alertmanager        2024.2           9b46a2a48036   23 hours ago        409MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor            2024.2           aac9a35ae18a   23 hours ago        363MB",
                                "registry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   5 weeks ago         11.5MB",
                                "registry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   2 months ago        334MB",
                                "registry.osism.tech/dockerhub/library/redis              7.4.7-alpine     13105d2858de   2 months ago        41.4MB",
                                "phpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   3 months ago        742MB",
                                "registry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   4 months ago        275MB",
                                "registry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   5 months ago        226MB",
                                "registry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   19 months ago       146MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-0",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "ea188795c6c1   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "2380e2574883   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "3ae4641f6e4e   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "070dbd4cc6dd   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "94284355a770   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "8d4471753615   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "462cbafa5532   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "7c54a847b027   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api",
                                "d48ceb80e386   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "d6305862ef7c   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana",
                                "62f14fda8460   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "239f51dbfcd3   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "ed57ad60b5e1   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "a0c26107ea02   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api",
                                "c62c68d72bdd   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "04766d8c4ed4   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "220256b5ba6e   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor",
                                "657a8dd50a44   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter",
                                "dc26c4b469d8   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter",
                                "741d2a945ecf   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter",
                                "d034b962119d   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor",
                                "61a3551be619   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api",
                                "b45c0ccd8ffa   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "8ec28903b5ee   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker",
                                "db903281ba06   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "ba58d1f35e07   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "df783840ea8c   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central",
                                "ff822d10bde2   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             placement_api",
                                "4af526a74eb1   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_api",
                                "9028ed0703eb   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "d0f93aa70de9   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-0",
                                "b011bf913b67   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker",
                                "b84cfe27a80b   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener",
                                "1b82559e5a07   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api",
                                "b17f16b43353   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "74106a3d1992   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "5bc9165738c3   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "e65e4adc51c6   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             horizon",
                                "622071ae0e63   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb",
                                "e58ff6e6c1bf   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards",
                                "a3d338fc1e63   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-0",
                                "77dcf89d1c62   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch",
                                "d06c84aed363   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "e3a535de601e   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "98cb6254ccb9   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "b0f455a2a0fd   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd",
                                "f056a300284a   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_sb_db",
                                "03ee67e9ef7c   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_nb_db",
                                "000f2103ee99   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-0",
                                "5583095e6fa6   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "8b7e7c971ce6   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             rabbitmq",
                                "e3997df8706a   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "1f42c21a1465   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "25af468da4ed   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db",
                                "0c9311ab9a35   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "cb2d99d448b9   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "39b51e8cb6bd   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "bc1c26491d65   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox",
                                "35ade5107c7a   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-0",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB",
                                "registry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB",
                                "registry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB",
                                "registry.osism.tech/kolla/ceilometer-notification             2024.2    56e7f91c3fab   23 hours ago   981MB",
                                "registry.osism.tech/kolla/ceilometer-central                  2024.2    e2d1fc7af8d7   23 hours ago   982MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB",
                                "registry.osism.tech/kolla/skyline-apiserver                   2024.2    1245eead9e1b   23 hours ago   995MB",
                                "registry.osism.tech/kolla/skyline-console                     2024.2    02f0b8255576   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/aodh-listener                       2024.2    a43a2e0ff586   23 hours ago   980MB",
                                "registry.osism.tech/kolla/aodh-notifier                       2024.2    464d439ccf30   23 hours ago   980MB",
                                "registry.osism.tech/kolla/aodh-evaluator                      2024.2    60e1bd484c4b   23 hours ago   980MB",
                                "registry.osism.tech/kolla/aodh-api                            2024.2    e620ea9760b8   23 hours ago   979MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-1",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-1'",
                                "",
                                "+ echo",
                                "+ osism container testbed-node-1 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "6d070a01e0ba   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "ef7abf888115   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "e19ed0666c5a   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "1f78bbd2edd7   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "ceaf14fb5ff3   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 4 minutes (healthy)              octavia_api",
                                "499ed120bece   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "b25e7cdd2871   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "7a0dc4bf0d1a   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes                        grafana",
                                "355e4bd73e99   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_api",
                                "a36617691ec6   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "2174ff3d2519   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "5483ab849afc   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "6743e3764bd4   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api",
                                "e727d5097483   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "7fd2e5b353eb   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "58b0950ef285   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "baa8dff29690   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor",
                                "6b1fd66c4011   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter",
                                "336eb7c80b7d   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter",
                                "ce571bae73c1   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter",
                                "8b13b6c200fb   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor",
                                "4852346278aa   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api",
                                "b9ae4948f091   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "4bfcaff86703   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker",
                                "9aed24fa714b   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "1195733f1647   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "00a03e326baf   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_central",
                                "99bf66434f3f   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api",
                                "0957003d948d   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api",
                                "438609e65c4d   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-1",
                                "340d4bad038e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "b674b86a5c23   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker",
                                "356e680e67cf   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener",
                                "2ac9b005a326   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api",
                                "e2e44ea2a969   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "03e8c3e57406   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon",
                                "f18cc30a0787   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "8ba0620e1cfb   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "c38a7f6e5bd3   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   20 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards",
                                "08e025b80e40   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb",
                                "fc97a97867f7   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch",
                                "0fae9b7a4be8   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-1",
                                "aea789e7bffb   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "dc3e7dc2688f   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "b1897adcd8c4   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "1b52b52981fa   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd",
                                "f6d9142811d9   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_sb_db",
                                "1d05f2e7c1ae   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_nb_db",
                                "28fc385ae1c1   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "84ac1be37239   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-1",
                                "b7e3fe639cae   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq",
                                "d3b0f0da827d   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   29 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "04ac4ac09c22   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "67fec7434b5d   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db",
                                "9bd790c0f9a2   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "243c917244b1   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "0c552c4f7a35   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "addc861f1229   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox",
                                "94d921799108   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-1",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB",
                                "registry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB",
                                "registry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "",
                                "## Containers @ testbed-node-2",
                                "",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "a81c4b68fe98   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "c86f5574a257   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "821a9d333f94   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_health_manager",
                                "00bd8e2743a0   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes                        octavia_driver_agent",
                                "efb72eae51a6   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "fff5f09e2d33   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "75b6386ec898   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "170195c7e378   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes                        grafana",
                                "5ce179670754   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api",
                                "e9fa2a8c6561   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "5a0070b70b0e   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_backup",
                                "eed660f315c7   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             cinder_volume",
                                "d4ebfecf6db8   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             glance_api",
                                "e0d2df62cfc0   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "eadbd876d04a   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "642cbc96058d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       prometheus_elasticsearch_exporter",
                                "c007ac101a92   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor",
                                "ca7cef65ce90   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_memcached_exporter",
                                "759c3a94e90c   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_mysqld_exporter",
                                "42f151ffaa33   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_node_exporter",
                                "4f47f34f6acb   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_conductor",
                                "9819d4c0621d   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             magnum_api",
                                "9ba1c3cfc4e0   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes (healthy)             neutron_server",
                                "81bb6c51b709   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_worker",
                                "a7cfb21e0460   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_mdns",
                                "677a1f3fdd63   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             designate_producer",
                                "bb326272f15a   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central",
                                "3dfa91430cf3   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api",
                                "42c3971145b0   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api",
                                "f678a7188e7c   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   15 minutes ago   Up 15 minutes                       ceph-mgr-testbed-node-2",
                                "420acd86d14e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "a9d4a3c52d07   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker",
                                "29cbcaa8bf53   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener",
                                "94b800366ebb   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_api",
                                "aae1d7a3322a   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "454a7bc660a9   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon",
                                "d5c0029d8b42   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "d0e831db5c33   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "99beef20bac9   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   21 minutes ago   Up 20 minutes (healthy)             opensearch_dashboards",
                                "3010d9f01137   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb",
                                "c8aa6760e44c   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch",
                                "b5b39c44de9e   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    22 minutes ago   Up 22 minutes                       ceph-crash-testbed-node-2",
                                "e5ddb7f2fe7f   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "8c345e3d480f   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "6ab1ede542d6   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "01fc0f6f50b2   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_northd",
                                "ec87d6288525   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_sb_db",
                                "54acb50e9e5d   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   27 minutes ago   Up 26 minutes                       ovn_nb_db",
                                "fb40cfc2dc34   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             rabbitmq",
                                "a69ff4dd3c75   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "01f2672e35a4   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-2",
                                "07229e97d6f6   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   29 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "5fb37a845f67   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "655e2208ef2e   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db",
                                "517d5e35db12   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "2e9774371971   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "c3e66589d1b9   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "39a23c099af8   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox",
                                "b0f8a3512151   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-2",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      8fb293e8afdf   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/grafana                             2024.2    abdd8fb3d2fa   23 hours ago   1.02GB",
                                "registry.osism.tech/kolla/cron                                2024.2    18397a23bf8e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    df276d2b69ca   23 hours ago   282MB",
                                "registry.osism.tech/kolla/memcached                           2024.2    ac6c32a007ae   23 hours ago   272MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    c93da708442d   23 hours ago   279MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    57c3a6a21e67   23 hours ago   417MB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    b58edbb0aebf   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    f47e7f5ea7ea   23 hours ago   1.53GB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    701b7d6baaed   23 hours ago   328MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    41bede1e920d   23 hours ago   675MB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    7623cc118b47   23 hours ago   585MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    51338ce4d7d7   23 hours ago   1.16GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    d62c246bdd51   23 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    3802479789de   23 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    8be0abeada49   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    3659f0b3baf8   23 hours ago   278MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    296069031ebe   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    39f19aba075b   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    a99e23a01ef0   23 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    18d09bc40715   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    aac9a35ae18a   23 hours ago   363MB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    51f8ffe4ca6e   23 hours ago   458MB",
                                "registry.osism.tech/kolla/keystone                            2024.2    60fd465cc253   23 hours ago   1.09GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    4528c8236356   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    e06e3ea5a6dd   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    f3982436dcf0   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    53fbf5812506   23 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    68c231276e30   23 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    bf190a1bf98f   23 hours ago   1.41GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    a59addf52e55   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    814d90987a6f   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    5ad6b55b48f5   23 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    4c252f4aac36   23 hours ago   1.37GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    418791135695   23 hours ago   981MB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    272846600dc6   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    1283503f2d5d   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    69111ffbeefe   23 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    9c00e76773d5   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    ce8f95c27da6   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    4d2bc901d4e5   23 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    d96c5aaa22bc   23 hours ago   1.25GB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    35d6a1495d78   23 hours ago   1.17GB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    7bcfdc7255dd   23 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    c916d4869f2e   23 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2de34a8fe642   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    bdd612831e36   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    2e6c856929d9   23 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    89d7df329e57   23 hours ago   994MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    dbc6b3f69c5d   23 hours ago   997MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9a0da1090b9d   23 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    e83ecb5b3f20   23 hours ago   997MB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    b64cd2c2f9fc   23 hours ago   1.1GB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    6b3dfa3e1db7   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    954866669b8c   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    8724a678703f   23 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    2e1bcc6f62cb   23 hours ago   846MB",
                                "+ sh -c /opt/configuration/scripts/check-services.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# Ceph status",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# Ceph status'",
                                "+ echo",
                                "+ ceph -s",
                                "  cluster:",
                                "    id:     11111111-1111-1111-1111-111111111111",
                                "    health: HEALTH_OK",
                                " ",
                                "  services:",
                                "    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 27m)",
                                "    mgr: testbed-node-2(active, since 14m), standbys: testbed-node-1, testbed-node-0",
                                "    mds: 1/1 daemons up, 2 standby",
                                "    osd: 6 osds: 6 up (since 23m), 6 in (since 24m)",
                                "    rgw: 3 daemons active (3 hosts, 1 zones)",
                                " ",
                                "  data:",
                                "    volumes: 1/1 healthy",
                                "    pools:   14 pools, 401 pgs",
                                "    objects: 552 objects, 2.2 GiB",
                                "    usage:   7.1 GiB used, 113 GiB / 120 GiB avail",
                                "    pgs:     401 active+clean",
                                " ",
                                "",
                                "# Ceph versions",
                                "",
                                "+ echo",
                                "+ echo '# Ceph versions'",
                                "+ echo",
                                "+ ceph versions",
                                "{",
                                "    \"mon\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"mgr\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"osd\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6",
                                "    },",
                                "    \"mds\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"rgw\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"overall\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18",
                                "    }",
                                "}",
                                "",
                                "# Ceph OSD tree",
                                "",
                                "+ echo",
                                "+ echo '# Ceph OSD tree'",
                                "+ echo",
                                "+ ceph osd df tree",
                                "ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              ",
                                "-1         0.11691         -  120 GiB  7.1 GiB  6.7 GiB    6 KiB  430 MiB  113 GiB  5.92  1.00    -          root default           ",
                                "-3         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-3",
                                " 0    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.5 GiB    1 KiB   70 MiB   18 GiB  7.70  1.30  200      up          osd.0          ",
                                " 4    hdd  0.01949   1.00000   20 GiB  844 MiB  771 MiB    1 KiB   74 MiB   19 GiB  4.13  0.70  190      up          osd.4          ",
                                "-5         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4",
                                " 2    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.4 GiB    1 KiB   74 MiB   19 GiB  7.38  1.25  206      up          osd.2          ",
                                " 5    hdd  0.01949   1.00000   20 GiB  912 MiB  843 MiB    1 KiB   70 MiB   19 GiB  4.46  0.75  186      up          osd.5          ",
                                "-7         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-5",
                                " 1    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.3 GiB    1 KiB   74 MiB   19 GiB  7.10  1.20  184      up          osd.1          ",
                                " 3    hdd  0.01949   1.00000   20 GiB  969 MiB  899 MiB    1 KiB   70 MiB   19 GiB  4.74  0.80  204      up          osd.3          ",
                                "                       TOTAL  120 GiB  7.1 GiB  6.7 GiB  9.3 KiB  430 MiB  113 GiB  5.92                                            ",
                                "MIN/MAX VAR: 0.70/1.30  STDDEV: 1.50",
                                "",
                                "# Ceph monitor status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph monitor status'",
                                "+ echo",
                                "+ ceph mon stat",
                                "e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2",
                                "",
                                "# Ceph quorum status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph quorum status'",
                                "+ echo",
                                "+ ceph quorum_status",
                                "+ jq",
                                "{",
                                "  \"election_epoch\": 8,",
                                "  \"quorum\": [",
                                "    0,",
                                "    1,",
                                "    2",
                                "  ],",
                                "  \"quorum_names\": [",
                                "    \"testbed-node-0\",",
                                "    \"testbed-node-1\",",
                                "    \"testbed-node-2\"",
                                "  ],",
                                "  \"quorum_leader_name\": \"testbed-node-0\",",
                                "  \"quorum_age\": 1656,",
                                "  \"features\": {",
                                "    \"quorum_con\": \"4540138322906710015\",",
                                "    \"quorum_mon\": [",
                                "      \"kraken\",",
                                "      \"luminous\",",
                                "      \"mimic\",",
                                "      \"osdmap-prune\",",
                                "      \"nautilus\",",
                                "      \"octopus\",",
                                "      \"pacific\",",
                                "      \"elector-pinging\",",
                                "      \"quincy\",",
                                "      \"reef\"",
                                "    ]",
                                "  },",
                                "  \"monmap\": {",
                                "    \"epoch\": 1,",
                                "    \"fsid\": \"11111111-1111-1111-1111-111111111111\",",
                                "    \"modified\": \"2026-01-13T00:47:57.735690Z\",",
                                "    \"created\": \"2026-01-13T00:47:57.735690Z\",",
                                "    \"min_mon_release\": 18,",
                                "    \"min_mon_release_name\": \"reef\",",
                                "    \"election_strategy\": 1,",
                                "    \"disallowed_leaders: \": \"\",",
                                "    \"stretch_mode\": false,",
                                "    \"tiebreaker_mon\": \"\",",
                                "    \"removed_ranks: \": \"\",",
                                "    \"features\": {",
                                "      \"persistent\": [",
                                "        \"kraken\",",
                                "        \"luminous\",",
                                "        \"mimic\",",
                                "        \"osdmap-prune\",",
                                "        \"nautilus\",",
                                "        \"octopus\",",
                                "        \"pacific\",",
                                "        \"elector-pinging\",",
                                "        \"quincy\",",
                                "        \"reef\"",
                                "      ],",
                                "      \"optional\": []",
                                "    },",
                                "    \"mons\": [",
                                "      {",
                                "        \"rank\": 0,",
                                "        \"name\": \"testbed-node-0\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.10:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.10:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.10:6789/0\",",
                                "        \"public_addr\": \"192.168.16.10:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 1,",
                                "        \"name\": \"testbed-node-1\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.11:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.11:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.11:6789/0\",",
                                "        \"public_addr\": \"192.168.16.11:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 2,",
                                "        \"name\": \"testbed-node-2\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.12:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.12:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.12:6789/0\",",
                                "        \"public_addr\": \"192.168.16.12:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      }",
                                "    ]",
                                "  }",
                                "}",
                                "",
                                "# Ceph free space status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph free space status'",
                                "+ echo",
                                "+ ceph df",
                                "--- RAW STORAGE ---",
                                "CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED",
                                "hdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92",
                                "TOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92",
                                " ",
                                "--- POOLS ---",
                                "POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL",
                                ".mgr                        1    1  577 KiB        2  1.1 MiB      0     52 GiB",
                                "cephfs_data                 2   32      0 B        0      0 B      0     35 GiB",
                                "cephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB",
                                "default.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.control         6   32      0 B        8      0 B      0     35 GiB",
                                "default.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB",
                                "default.rgw.meta            8   32      0 B        0      0 B      0     35 GiB",
                                ".rgw.root                   9   32  1.4 KiB        4   32 KiB      0     52 GiB",
                                "backups                    10   32     19 B        2   12 KiB      0     35 GiB",
                                "volumes                    11   32     19 B        2   12 KiB      0     35 GiB",
                                "images                     12   32  2.2 GiB      299  6.7 GiB   5.98     35 GiB",
                                "metrics                    13   32     19 B        2   12 KiB      0     35 GiB",
                                "vms                        14   32     19 B        2   12 KiB      0     35 GiB",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ ! -e /etc/redhat-release ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-01-13 01:16:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a66d0131-6367-4bd5-a2fd-6d1bec74496b (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 01:16:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a66d0131-6367-4bd5-a2fd-6d1bec74496b (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 13 January 2026  01:16:11 +0000 (0:00:00.266)       0:00:00.266 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 13 January 2026  01:16:12 +0000 (0:00:01.377)       0:00:01.644 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 13 January 2026  01:16:13 +0000 (0:00:01.415)       0:00:03.059 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 13 January 2026  01:16:21 +0000 (0:00:07.560)       0:00:10.620 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:16:22 +0000 (0:00:00.545)       0:00:11.166 ******* ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 7.56s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.42s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.38s",
                                "Gather facts for all hosts ---------------------------------------------- 0.55s",
                                "+ osism validate ceph-mons",
                                "",
                                "PLAY [Ceph validate mons] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 13 January 2026  01:16:39 +0000 (0:00:00.421)       0:00:00.421 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 13 January 2026  01:16:40 +0000 (0:00:00.777)       0:00:01.198 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 13 January 2026  01:16:41 +0000 (0:00:00.933)       0:00:02.131 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Tuesday 13 January 2026  01:16:41 +0000 (0:00:00.129)       0:00:02.261 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Tuesday 13 January 2026  01:16:41 +0000 (0:00:00.279)       0:00:02.540 ******* ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Tuesday 13 January 2026  01:16:42 +0000 (0:00:00.975)       0:00:03.515 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Tuesday 13 January 2026  01:16:43 +0000 (0:00:00.288)       0:00:03.804 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:16:43 +0000 (0:00:00.485)       0:00:04.289 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mon is not running] ********************",
                                "Tuesday 13 January 2026  01:16:44 +0000 (0:00:00.315)       0:00:04.605 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mon is running] ************************",
                                "Tuesday 13 January 2026  01:16:44 +0000 (0:00:00.286)       0:00:04.891 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:16:44 +0000 (0:00:00.511)       0:00:05.403 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.263)       0:00:05.667 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.245)       0:00:05.912 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.234)       0:00:06.147 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.068)       0:00:06.215 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.068)       0:00:06.284 ******* ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:16:45 +0000 (0:00:00.074)       0:00:06.359 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Tuesday 13 January 2026  01:16:46 +0000 (0:00:00.232)       0:00:06.592 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare quorum test vars] ************************************************",
                                "Tuesday 13 January 2026  01:16:46 +0000 (0:00:00.257)       0:00:06.850 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get monmap info from one mon container] **********************************",
                                "Tuesday 13 January 2026  01:16:46 +0000 (0:00:00.133)       0:00:06.983 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set quorum test data] ****************************************************",
                                "Tuesday 13 January 2026  01:16:48 +0000 (0:00:01.696)       0:00:08.680 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail quorum test if not all monitors are in quorum] **********************",
                                "Tuesday 13 January 2026  01:16:48 +0000 (0:00:00.519)       0:00:09.200 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass quorum test if all monitors are in quorum] **************************",
                                "Tuesday 13 January 2026  01:16:48 +0000 (0:00:00.133)       0:00:09.333 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set fsid test vars] ******************************************************",
                                "Tuesday 13 January 2026  01:16:49 +0000 (0:00:00.324)       0:00:09.658 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail Cluster FSID test if FSID does not match configuration] *************",
                                "Tuesday 13 January 2026  01:16:49 +0000 (0:00:00.300)       0:00:09.958 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass Cluster FSID test if it matches configuration] **********************",
                                "Tuesday 13 January 2026  01:16:49 +0000 (0:00:00.108)       0:00:10.067 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare status test vars] ************************************************",
                                "Tuesday 13 January 2026  01:16:49 +0000 (0:00:00.126)       0:00:10.193 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather status data] ******************************************************",
                                "Tuesday 13 January 2026  01:16:49 +0000 (0:00:00.108)       0:00:10.301 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set health test data] ****************************************************",
                                "Tuesday 13 January 2026  01:16:50 +0000 (0:00:01.243)       0:00:11.544 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable] *************************",
                                "Tuesday 13 January 2026  01:16:51 +0000 (0:00:00.298)       0:00:11.843 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if health is acceptable] *****************************",
                                "Tuesday 13 January 2026  01:16:51 +0000 (0:00:00.153)       0:00:11.997 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable (strict)] ****************",
                                "Tuesday 13 January 2026  01:16:51 +0000 (0:00:00.145)       0:00:12.143 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if status is OK (strict)] ****************************",
                                "Tuesday 13 January 2026  01:16:51 +0000 (0:00:00.326)       0:00:12.470 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 13 January 2026  01:16:52 +0000 (0:00:00.149)       0:00:12.619 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 13 January 2026  01:16:52 +0000 (0:00:00.239)       0:00:12.859 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:16:52 +0000 (0:00:00.234)       0:00:13.093 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:16:54 +0000 (0:00:01.750)       0:00:14.843 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:16:54 +0000 (0:00:00.268)       0:00:15.112 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:54 +0000 (0:00:00.258)       0:00:15.370 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:54 +0000 (0:00:00.069)       0:00:15.439 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:16:54 +0000 (0:00:00.084)       0:00:15.523 ******* ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 13 January 2026  01:16:55 +0000 (0:00:00.093)       0:00:15.617 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:16:56 +0000 (0:00:01.482)       0:00:17.099 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-01-13T01:16:40+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:16:57 +0000 (0:00:00.827)       0:00:17.926 ******* ",
                                "=============================================================================== ",
                                "Aggregate test results step one ----------------------------------------- 1.75s",
                                "Get monmap info from one mon container ---------------------------------- 1.70s",
                                "Write report file ------------------------------------------------------- 1.48s",
                                "Gather status data ------------------------------------------------------ 1.24s",
                                "Get container info ------------------------------------------------------ 0.98s",
                                "Create report output directory ------------------------------------------ 0.93s",
                                "Print report file information ------------------------------------------- 0.83s",
                                "Get timestamp for report file ------------------------------------------- 0.78s",
                                "Set quorum test data ---------------------------------------------------- 0.52s",
                                "Set test result to passed if ceph-mon is running ------------------------ 0.51s",
                                "Set test result to passed if container is existing ---------------------- 0.49s",
                                "Fail cluster-health if health is not acceptable (strict) ---------------- 0.33s",
                                "Pass quorum test if all monitors are in quorum -------------------------- 0.32s",
                                "Prepare test data ------------------------------------------------------- 0.32s",
                                "Set fsid test vars ------------------------------------------------------ 0.30s",
                                "Set health test data ---------------------------------------------------- 0.30s",
                                "Set test result to failed if container is missing ----------------------- 0.29s",
                                "Set test result to failed if ceph-mon is not running -------------------- 0.29s",
                                "Prepare test data for container existance test -------------------------- 0.28s",
                                "Aggregate test results step two ----------------------------------------- 0.27s",
                                "+ osism validate ceph-mgrs",
                                "",
                                "PLAY [Ceph validate mgrs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 13 January 2026  01:17:09 +0000 (0:00:00.432)       0:00:00.432 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 13 January 2026  01:17:10 +0000 (0:00:00.792)       0:00:01.225 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 13 January 2026  01:17:11 +0000 (0:00:01.036)       0:00:02.261 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Tuesday 13 January 2026  01:17:11 +0000 (0:00:00.130)       0:00:02.392 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Tuesday 13 January 2026  01:17:11 +0000 (0:00:00.286)       0:00:02.678 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Tuesday 13 January 2026  01:17:13 +0000 (0:00:01.251)       0:00:03.930 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Tuesday 13 January 2026  01:17:13 +0000 (0:00:00.273)       0:00:04.203 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:13 +0000 (0:00:00.471)       0:00:04.675 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mgr is not running] ********************",
                                "Tuesday 13 January 2026  01:17:14 +0000 (0:00:00.295)       0:00:04.970 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mgr is running] ************************",
                                "Tuesday 13 January 2026  01:17:14 +0000 (0:00:00.282)       0:00:05.252 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:17:14 +0000 (0:00:00.494)       0:00:05.747 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.238)       0:00:05.985 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.284)       0:00:06.270 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.251)       0:00:06.521 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.071)       0:00:06.593 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.070)       0:00:06.663 ******* ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:17:15 +0000 (0:00:00.076)       0:00:06.740 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Tuesday 13 January 2026  01:17:16 +0000 (0:00:00.245)       0:00:06.986 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Define mgr module test vars] *********************************************",
                                "Tuesday 13 January 2026  01:17:16 +0000 (0:00:00.253)       0:00:07.239 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather list of mgr modules] **********************************************",
                                "Tuesday 13 January 2026  01:17:16 +0000 (0:00:00.115)       0:00:07.355 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Parse mgr module list from json] *****************************************",
                                "Tuesday 13 January 2026  01:17:18 +0000 (0:00:01.834)       0:00:09.189 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Extract list of enabled mgr modules] *************************************",
                                "Tuesday 13 January 2026  01:17:18 +0000 (0:00:00.427)       0:00:09.616 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail test if mgr modules are disabled that should be enabled] ************",
                                "Tuesday 13 January 2026  01:17:19 +0000 (0:00:00.327)       0:00:09.944 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass test if required mgr modules are enabled] ***************************",
                                "Tuesday 13 January 2026  01:17:19 +0000 (0:00:00.124)       0:00:10.069 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 13 January 2026  01:17:19 +0000 (0:00:00.154)       0:00:10.223 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 13 January 2026  01:17:19 +0000 (0:00:00.249)       0:00:10.472 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:17:19 +0000 (0:00:00.232)       0:00:10.704 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:17:20 +0000 (0:00:01.211)       0:00:11.915 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:17:21 +0000 (0:00:00.267)       0:00:12.182 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:21 +0000 (0:00:00.239)       0:00:12.423 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:21 +0000 (0:00:00.068)       0:00:12.491 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:21 +0000 (0:00:00.067)       0:00:12.559 ******* ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 13 January 2026  01:17:21 +0000 (0:00:00.255)       0:00:12.815 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:17:23 +0000 (0:00:01.299)       0:00:14.115 ******* ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-01-13T01:17:10+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:17:23 +0000 (0:00:00.403)       0:00:14.518 ******* ",
                                "=============================================================================== ",
                                "Gather list of mgr modules ---------------------------------------------- 1.83s",
                                "Write report file ------------------------------------------------------- 1.30s",
                                "Get container info ------------------------------------------------------ 1.25s",
                                "Aggregate test results step one ----------------------------------------- 1.21s",
                                "Create report output directory ------------------------------------------ 1.04s",
                                "Get timestamp for report file ------------------------------------------- 0.79s",
                                "Set test result to passed if ceph-mgr is running ------------------------ 0.49s",
                                "Set test result to passed if container is existing ---------------------- 0.47s",
                                "Parse mgr module list from json ----------------------------------------- 0.43s",
                                "Print report file information ------------------------------------------- 0.40s",
                                "Flush handlers ---------------------------------------------------------- 0.39s",
                                "Extract list of enabled mgr modules ------------------------------------- 0.33s",
                                "Prepare test data ------------------------------------------------------- 0.30s",
                                "Prepare test data for container existance test -------------------------- 0.29s",
                                "Aggregate test results step two ----------------------------------------- 0.28s",
                                "Set test result to failed if ceph-mgr is not running -------------------- 0.28s",
                                "Set test result to failed if container is missing ----------------------- 0.27s",
                                "Aggregate test results step two ----------------------------------------- 0.27s",
                                "Fail due to missing containers ------------------------------------------ 0.25s",
                                "Aggregate test results step three --------------------------------------- 0.25s",
                                "+ osism validate ceph-osds",
                                "",
                                "PLAY [Ceph validate OSDs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 13 January 2026  01:17:40 +0000 (0:00:00.408)       0:00:00.408 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 13 January 2026  01:17:41 +0000 (0:00:00.808)       0:00:01.216 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 13 January 2026  01:17:42 +0000 (0:00:00.499)       0:00:01.716 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 13 January 2026  01:17:42 +0000 (0:00:00.669)       0:00:02.386 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Tuesday 13 January 2026  01:17:43 +0000 (0:00:00.117)       0:00:02.504 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Tuesday 13 January 2026  01:17:43 +0000 (0:00:00.130)       0:00:02.634 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Tuesday 13 January 2026  01:17:43 +0000 (0:00:00.301)       0:00:02.936 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Tuesday 13 January 2026  01:17:43 +0000 (0:00:00.142)       0:00:03.078 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate total number of OSDs in cluster] *******************************",
                                "Tuesday 13 January 2026  01:17:43 +0000 (0:00:00.312)       0:00:03.391 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:44 +0000 (0:00:00.601)       0:00:03.992 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers on host] *********************************",
                                "Tuesday 13 January 2026  01:17:45 +0000 (0:00:00.480)       0:00:04.472 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'd3a78fcb007e9c9cf8beb00c41c41182ff4f7141b070b40787de7ec0e5920240', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '74ca70dc77d6f055dc1977d50acf32ef5a65859a316ff150c7e956e8ce8f113e', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0f057698e43a39bb26c65a70895321e4c94977d93367f4a34044d6235e50a35d', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '501c2d37b004c04e960539846b79750a31df18d4984494bccb3d321f67b12355', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'fc81a65026c293243876f3a037c928c9e8302c714721c298c36dc8d11fef9f2c', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '16f8e6164b65c1962d79a58cc1df74729f2818837581bec4bb9fd32af3d46c0d', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '77c17e344c751f33b4e619cd49d30ffb165185980a41f698ee4acbf212a87236', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ac50de9f26bbe3e9c8ff977156ab573314956294e9d033aabfec97231c5de069', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'dbf500f7855dadb661620cb1f20e7a118a01cee114c3768e9b58a9de098a0951', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'dcc4c3b0e0ce368d05ffcfe9a38a6a63dbd36bfc0220a82a31c783b587b0d12b', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': '9ff012654489585a85c16e6a37432be319b40455fe8df31375d59993386344b0', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': 'add6cf9e8ba3561a450d02e573c16e4ec90b99d2969871fd5d8194a7d854fb6a', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8c034d9af798b7f0d7980314c29c6fa6f1c01c9ad236cf05c53057267ff850ae', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0476ed090495228a22c133fa90cf3661a6ce0a9edad3173504a702edd4a6f2e7', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'c24bd64d85708f35ec665c98850bf1b7a8be7d31dd6eb78d04f07ea294c03ebc', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e6956e22e953a73a77428962cadb3516942f8cbfd1c7f1e8b8bacd3e2e5d8896', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '0a389f80a55f73a57fc38c284bd5b4817cdefaac8ea27b5acb2311f16110e7a7', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ff37babf5b9ed1b3a668f1c5b88f9966d8266abd23565e787200476559530c29', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1de249b56223bcefe828864482a846e0ef2aed56db7e4b7fe0b3f301a8e6a6b6', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '203cae7e9471b4926a0ba5cca1e546eab5204e49a478a8d214ce0305c698ac24', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8ee1be56a0b59d70014e7aacbdd8c7b3e88ef9c11ac1a606f8c66be0478d4d24', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6641d8b49495231b8e1b3f96ca8ee6c0dc613562be2f4df89b8f1a5058b21511', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0eea71216deabba3ffac7a2f099e5eaddb2e82c95eeab999562b2cfaf89b061e', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c3c80c73b6070420d84aee4a793e8abf74aec8f5c2ba8b2ac3f84a94c546ed32', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ef7e548b1eb7fea6e0b310c2dec11236a0162d47a8f555e3963a86a4657d1f01', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c77125e3250d11076354289b7976a700b13354225538c4c521c2e1e903bf8e94', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3ab45e01af15fa611a229b1c189e995bce34019640c57b4b2f6a6cf1b16745c5', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'dd0d63e50b80ccb93e2755428dc7a9d03e7fbfd8406794efafa36fe04fd2d261', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '70a450169183dee9c5f573c57a313c5d7dc19d008fd64ce40f51cfacb51d2998', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': 'd7145a18aee01206e1044557c942e7512c26d36b8ea41dd3bd90dfa1375c494a', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '0cda99d5c6691e323e9278321577a2ef44fbe4cd1728ce9f9e010ec19aeb23e9', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e33387bfa156e99b01fde1badacac8ccb7c1e86de42dee06ae8f6248329f3c13', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '25010fc9c54f4efff56dd3d7d36bbbb6978a3d3b714a88e824d15b9670cab64e', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '296655f9405254e03e3e41af4b0c690c30cb7444585d0a8594379f81d663231b', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '20aee76fafcd6639019c80ed9e8e7410768973fbe6af6e91304b80ee51c6eb42', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '761948de6bc91d66b89370e7d945883288207c20eb969377c91dae379d7edd50', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b328d61fdf17323e257ebfd30e1bad731e307c21fecc881c0d2b0b7dbf9b598b', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '23ab7758342847ebd9fb62766ebc2485a7160fc866346d1eefa9cd7ee949d04f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '3ab1060c447cc5406468aef271af340002df988de2302a437d5bb4223e809690', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '0665162f52b763b5c76931e9af0ac5ddca90c79858155af311da5dec36ffae49', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '98b346020f62df488d4e839711b116281e3169623da33161f52db7bf96da431f', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 12 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '98dbddf37d2c7c907799c21d6de41884478c4765dce765b1cd20068da70f3006', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '4245043ad1a72dbc1b0fd2df2eaf5ca0ebdfb611e01db3906f638ed40bd9d3e0', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9e4bc82239fc81ff290bffbcf7402743ef3d5caacee719ffd53a7d64922b653f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '35a21964a68f180972a4493eee94d2297773e4641b4a8b5fc45aa524d89c0331', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 22 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1e933cbe1b7199bdb6c7dc8be19f3f068c99fe685ce5ca443ba0040fd5129de7', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '5b124b584b96f6e367fb9719c890f8fe2a7423c3b676041d3b94fe8c9864fac8', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '665fd7911952682cc2e6c6c1adc4e90293e5a58e0dc18cbdc6f89c6689f2ea75', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'dded726ea7cd20156756fac7346c2c27432e38b646e04072a7add8bd5f9bce38', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fea967d64cbb6e2e73642576dc14f4c869b4afcc8691d2d277c84322fe2d1d3a', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9c6af98f19ef3213321fa58afcf4a625c5bcb6557e92f5a59f835066551aa5a6', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '3890e7b35cfca588e566c42f4a3a874bf3e39e84e26e89b2a52a088438fbcfac', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8d0fd15f8d0ca6eda9ff73dab70c2224170befd946471b96881cf334b7b18763', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '14dfc3f860f2cfff4737b8970309e3f5d84a8d27141a4b67b7aeb9c653936983', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers on host] ********************************",
                                "Tuesday 13 January 2026  01:17:45 +0000 (0:00:00.480)       0:00:04.953 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed when count of containers is wrong] *************",
                                "Tuesday 13 January 2026  01:17:45 +0000 (0:00:00.301)       0:00:05.254 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if count matches] ******************************",
                                "Tuesday 13 January 2026  01:17:46 +0000 (0:00:00.494)       0:00:05.749 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:46 +0000 (0:00:00.286)       0:00:06.035 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers that are not running] ********************",
                                "Tuesday 13 January 2026  01:17:46 +0000 (0:00:00.276)       0:00:06.312 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers that are not running] *******************",
                                "Tuesday 13 January 2026  01:17:47 +0000 (0:00:00.300)       0:00:06.612 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Tuesday 13 January 2026  01:17:47 +0000 (0:00:00.473)       0:00:07.086 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Tuesday 13 January 2026  01:17:47 +0000 (0:00:00.294)       0:00:07.381 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if all containers are running] *****************",
                                "Tuesday 13 January 2026  01:17:48 +0000 (0:00:00.275)       0:00:07.656 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:17:48 +0000 (0:00:00.277)       0:00:07.933 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:17:48 +0000 (0:00:00.470)       0:00:08.404 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:17:49 +0000 (0:00:00.626)       0:00:09.030 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:49 +0000 (0:00:00.231)       0:00:09.261 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:49 +0000 (0:00:00.068)       0:00:09.329 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:17:49 +0000 (0:00:00.080)       0:00:09.410 ******* ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:17:50 +0000 (0:00:00.083)       0:00:09.493 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail early due to containers not running] ********************************",
                                "Tuesday 13 January 2026  01:17:50 +0000 (0:00:00.251)       0:00:09.744 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:50 +0000 (0:00:00.238)       0:00:09.983 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set _mon_hostname fact] **************************************************",
                                "Tuesday 13 January 2026  01:17:50 +0000 (0:00:00.290)       0:00:10.274 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get ceph osd tree] *******************************************************",
                                "Tuesday 13 January 2026  01:17:51 +0000 (0:00:00.239)       0:00:10.513 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Parse osd tree from JSON] ************************************************",
                                "Tuesday 13 January 2026  01:17:53 +0000 (0:00:02.142)       0:00:12.655 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get OSDs that are not up or in] ******************************************",
                                "Tuesday 13 January 2026  01:17:53 +0000 (0:00:00.132)       0:00:12.787 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail test if OSDs are not up or in] **************************************",
                                "Tuesday 13 January 2026  01:17:53 +0000 (0:00:00.289)       0:00:13.077 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Pass test if OSDs are all up and in] *************************************",
                                "Tuesday 13 January 2026  01:17:53 +0000 (0:00:00.121)       0:00:13.198 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:53 +0000 (0:00:00.108)       0:00:13.306 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [List ceph LVM volumes and collect data] **********************************",
                                "Tuesday 13 January 2026  01:17:54 +0000 (0:00:00.281)       0:00:13.588 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Parse LVM data as JSON] **************************************************",
                                "Tuesday 13 January 2026  01:17:56 +0000 (0:00:02.739)       0:00:16.328 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get unencrypted and encrypted OSDs] **************************************",
                                "Tuesday 13 January 2026  01:17:57 +0000 (0:00:00.484)       0:00:16.812 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of encrypted OSDs does not match] **************************",
                                "Tuesday 13 January 2026  01:17:57 +0000 (0:00:00.478)       0:00:17.291 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of encrypted OSDs equals count of OSDs] ********************",
                                "Tuesday 13 January 2026  01:17:58 +0000 (0:00:00.290)       0:00:17.581 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of unencrypted OSDs does not match] ************************",
                                "Tuesday 13 January 2026  01:17:58 +0000 (0:00:00.461)       0:00:18.043 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************",
                                "Tuesday 13 January 2026  01:17:58 +0000 (0:00:00.276)       0:00:18.319 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 13 January 2026  01:17:59 +0000 (0:00:00.263)       0:00:18.583 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get CRUSH node data of each OSD host and root node childs] ***************",
                                "Tuesday 13 January 2026  01:17:59 +0000 (0:00:00.547)       0:00:19.130 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate sub test expression results] ***********************************",
                                "Tuesday 13 January 2026  01:18:00 +0000 (0:00:00.903)       0:00:20.034 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail test if any sub test failed] ****************************************",
                                "Tuesday 13 January 2026  01:18:00 +0000 (0:00:00.296)       0:00:20.331 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass test if no sub test failed] *****************************************",
                                "Tuesday 13 January 2026  01:18:01 +0000 (0:00:00.268)       0:00:20.599 ******* ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 13 January 2026  01:18:01 +0000 (0:00:00.287)       0:00:20.887 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 13 January 2026  01:18:01 +0000 (0:00:00.234)       0:00:21.122 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 13 January 2026  01:18:02 +0000 (0:00:00.694)       0:00:21.816 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 13 January 2026  01:18:03 +0000 (0:00:01.512)       0:00:23.329 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 13 January 2026  01:18:04 +0000 (0:00:00.278)       0:00:23.607 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:18:04 +0000 (0:00:00.230)       0:00:23.838 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:18:04 +0000 (0:00:00.068)       0:00:23.906 ******* ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 13 January 2026  01:18:04 +0000 (0:00:00.066)       0:00:23.972 ******* ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 13 January 2026  01:18:04 +0000 (0:00:00.069)       0:00:24.041 ******* ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 13 January 2026  01:18:05 +0000 (0:00:01.240)       0:00:25.282 ******* ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-01-13T01:17:41+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:18:06 +0000 (0:00:00.386)       0:00:25.668 ******* ",
                                "=============================================================================== ",
                                "List ceph LVM volumes and collect data ---------------------------------- 2.74s",
                                "Get ceph osd tree ------------------------------------------------------- 2.14s",
                                "Aggregate test results step one ----------------------------------------- 1.51s",
                                "Write report file ------------------------------------------------------- 1.24s",
                                "Get CRUSH node data of each OSD host and root node childs --------------- 0.90s",
                                "Get timestamp for report file ------------------------------------------- 0.81s",
                                "Set validation result to failed if a test failed ------------------------ 0.69s",
                                "Create report output directory ------------------------------------------ 0.67s",
                                "Aggregate test results step two ----------------------------------------- 0.63s",
                                "Calculate total number of OSDs in cluster ------------------------------- 0.60s",
                                "Prepare test data ------------------------------------------------------- 0.55s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.50s",
                                "Set test result to failed when count of containers is wrong ------------- 0.49s",
                                "Parse LVM data as JSON -------------------------------------------------- 0.48s",
                                "Prepare test data ------------------------------------------------------- 0.48s",
                                "Get list of ceph-osd containers on host --------------------------------- 0.48s",
                                "Get unencrypted and encrypted OSDs -------------------------------------- 0.48s",
                                "Get count of ceph-osd containers that are not running ------------------- 0.47s",
                                "Aggregate test results step one ----------------------------------------- 0.47s",
                                "Pass if count of encrypted OSDs equals count of OSDs -------------------- 0.46s",
                                "+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.193.234",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ source /etc/os-release",
                                "++ PRETTY_NAME='Ubuntu 24.04.3 LTS'",
                                "++ NAME=Ubuntu",
                                "++ VERSION_ID=24.04",
                                "++ VERSION='24.04.3 LTS (Noble Numbat)'",
                                "++ VERSION_CODENAME=noble",
                                "++ ID=ubuntu",
                                "++ ID_LIKE=debian",
                                "++ HOME_URL=https://www.ubuntu.com/",
                                "++ SUPPORT_URL=https://help.ubuntu.com/",
                                "++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/",
                                "++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy",
                                "++ UBUNTU_CODENAME=noble",
                                "++ LOGO=ubuntu-logo",
                                "+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]",
                                "+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'",
                                "+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "",
                                "# Status of Elasticsearch",
                                "",
                                "+ pushd /opt/configuration/contrib",
                                "+ echo",
                                "+ echo '# Status of Elasticsearch'",
                                "+ echo",
                                "+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s",
                                "OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0",
                                "",
                                "+ echo",
                                "+ echo '# Status of MariaDB'",
                                "# Status of MariaDB",
                                "",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status database",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: 4c7e25da-f01a-11f0-8988-c257ed0aca95\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.24(ra6b53429)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: 81b21e74-f01a-11f0-9b30-e2d7a1171da6\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000127307\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 5073 local commits, 7788 replicated, 122 received\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 23 minutes, 23 seconds\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 128 connected, 1 running\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 137868 total, 0 slow\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 42\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Prometheus",
                                "",
                                "+ echo",
                                "+ echo '# Status of Prometheus'",
                                "+ echo",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy",
                                "Unauthorized",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready",
                                "Unauthorized",
                                "",
                                "# Status of RabbitMQ",
                                "",
                                "+ echo",
                                "+ echo '# Status of RabbitMQ'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status messaging",
                                "\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 204, Channels: 203, Queues: 173\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 119/1024\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 71/832\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 204, Channels: 203, Queues: 173\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.8 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 124/1024\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 78/832\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 204, Channels: 203, Queues: 173\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 7.0/s publish, 7.6/s deliver\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.9 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 101/1024\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 55/832\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Redis",
                                "",
                                "+ echo",
                                "+ echo '# Status of Redis'",
                                "+ echo",
                                "+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j",
                                "TCP OK - 0.001 second response time on 192.168.16.10 port 6379|time=0.001242s;;;0.000000;10.000000",
                                "+ popd",
                                "",
                                "# Create backup of MariaDB database",
                                "",
                                "+ echo",
                                "+ echo '# Create backup of MariaDB database'",
                                "+ echo",
                                "+ osism apply mariadb_backup -e mariadb_backup_type=full",
                                "\u001b[32m2026-01-13 01:18:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 062ddd45-8a86-4a8a-9d97-879af9c04429 (mariadb_backup) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 01:18:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 062ddd45-8a86-4a8a-9d97-879af9c04429 (mariadb_backup) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 13 January 2026  01:19:00 +0000 (0:00:00.168)       0:00:00.168 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 13 January 2026  01:19:00 +0000 (0:00:00.321)       0:00:00.490 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Tuesday 13 January 2026  01:19:00 +0000 (0:00:00.566)       0:00:01.057 ******* ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 13 January 2026  01:19:01 +0000 (0:00:00.413)       0:00:01.470 ******* ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB container facts] ***********************************",
                                "Tuesday 13 January 2026  01:19:01 +0000 (0:00:00.541)       0:00:02.012 ******* ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Taking full database backup via Mariabackup] *******************",
                                "Tuesday 13 January 2026  01:19:05 +0000 (0:00:03.276)       0:00:05.289 ******* ",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Tuesday 13 January 2026  01:19:46 +0000 (0:00:41.685)       0:00:46.974 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Tuesday 13 January 2026  01:19:47 +0000 (0:00:00.299)       0:00:47.274 ******* ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:19:47 +0000 (0:00:00.393)       0:00:47.668 ******* ",
                                "=============================================================================== ",
                                "mariadb : Taking full database backup via Mariabackup ------------------ 41.69s",
                                "mariadb : Get MariaDB container facts ----------------------------------- 3.28s",
                                "Group hosts based on enabled services ----------------------------------- 0.57s",
                                "mariadb : include_tasks ------------------------------------------------- 0.54s",
                                "mariadb : Group MariaDB hosts based on shards --------------------------- 0.41s",
                                "Include mariadb post-upgrade.yml ---------------------------------------- 0.39s",
                                "Group hosts based on Kolla action --------------------------------------- 0.32s",
                                "Include mariadb post-deploy.yml ----------------------------------------- 0.30s",
                                "+ sh -c /opt/configuration/scripts/check/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# OpenStack endpoints",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ export OS_CLOUD=admin",
                                "+ OS_CLOUD=admin",
                                "+ echo",
                                "+ echo '# OpenStack endpoints'",
                                "+ echo",
                                "+ openstack endpoint list",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| 06574ad6874b4068b9bc0715f33ce4d8 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |",
                                "| 19208f7c81644839bb2114dd1b2aa712 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |",
                                "| 256c014a163d4bb3b0a21d6d415e3f1f | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |",
                                "| 27ffb9623e47436c9a7b873bfe49df78 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |",
                                "| 2938964b65df47a2b60afc351c312186 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |",
                                "| 3556e304fef940c7ac51bba018c10a60 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |",
                                "| 462fe6fc292349acba4e5368c004e13f | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |",
                                "| 582285c0b7354a218609896e63a7c1ca | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |",
                                "| 61be64e294ac46e487d30d3c3dbc5a80 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |",
                                "| 620a43c623624295a911aabcdc22cff3 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |",
                                "| 87a2354dffd945359975254d4d61b7a0 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |",
                                "| 896f8523f4b14bf59f1dae3a360b1340 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |",
                                "| 9d0b4c8fe69f4495be961f98e0af2b7a | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |",
                                "| b437ac5f277d43388ed5f54aca4d5319 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |",
                                "| b4af2bdf24e9452293a735ab95ecb545 | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |",
                                "| bb9a954b184546e4a6cbfc27a68b076a | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |",
                                "| c5df2317bce0469fb96e3c128fa439a3 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |",
                                "| c7487228917044509d34aef0ae5e2e99 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |",
                                "| d5e461332ec449ed9e1a8039c62648d1 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |",
                                "| dab0159c02e54dd38966f555de81adab | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |",
                                "| df150eb0270d4d5c97a7ce8e4aac8bb7 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |",
                                "| f1d61a022aaf45de8479808af7cbc2c0 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "",
                                "# Cinder",
                                "",
                                "+ echo",
                                "+ echo '# Cinder'",
                                "+ echo",
                                "+ openstack volume service list",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| Binary           | Host                       | Zone     | Status  | State | Updated At                 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |",
                                "| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |",
                                "| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-01-13T01:19:49.000000 |",
                                "| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:49.000000 |",
                                "| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:51.000000 |",
                                "| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-01-13T01:19:52.000000 |",
                                "| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-01-13T01:19:44.000000 |",
                                "| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-01-13T01:19:45.000000 |",
                                "| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-01-13T01:19:46.000000 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "",
                                "# Neutron",
                                "",
                                "+ echo",
                                "+ echo '# Neutron'",
                                "+ echo",
                                "+ openstack network agent list",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |",
                                "| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "+ openstack network service provider list",
                                "+---------------+------+---------+",
                                "| Service Type  | Name | Default |",
                                "+---------------+------+---------+",
                                "| L3_ROUTER_NAT | ovn  | True    |",
                                "+---------------+------+---------+",
                                "",
                                "# Nova",
                                "",
                                "+ echo",
                                "+ echo '# Nova'",
                                "+ echo",
                                "+ openstack compute service list",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ad6e1ef3-4b1f-47bb-a5e8-f20d1f9715e8 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-01-13T01:19:55.000000 |",
                                "| 02ecf3a3-9236-4c9e-a41f-4511b5e69fae | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-01-13T01:19:56.000000 |",
                                "| e20dd9f4-0b2d-4012-8452-dcbf772284e2 | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-01-13T01:19:57.000000 |",
                                "| 6571e646-fa52-46f1-a7ee-4c48984f15e4 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-01-13T01:19:59.000000 |",
                                "| 5aeeec1b-e06e-4ab7-bcd5-3e7efb785f72 | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-01-13T01:20:00.000000 |",
                                "| 3f0b0182-a9a1-42f3-b8ed-0f2b28e43402 | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-01-13T01:20:00.000000 |",
                                "| 29e50209-78a5-4012-a38a-bca2851525b0 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-01-13T01:20:00.000000 |",
                                "| bb892313-8417-4ff5-8c2f-f9bf9aa0b005 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-01-13T01:20:00.000000 |",
                                "| 3d63b899-2787-417e-96ce-91b3e6f96fad | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-01-13T01:20:01.000000 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "+ openstack hypervisor list",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| 8f93afd9-4cc1-49f6-82ec-3a6527093679 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |",
                                "| cbb720b7-82d5-452e-bf67-987b544a2ead | testbed-node-3      | QEMU            | 192.168.16.13 | up    |",
                                "| 1269b2e9-7f60-43bd-a2a9-9cd7fb028269 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "+ echo",
                                "",
                                "# Run OpenStack test play",
                                "",
                                "+ echo '# Run OpenStack test play'",
                                "+ echo",
                                "+ osism apply --environment openstack test",
                                "\u001b[32m2026-01-13 01:20:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m",
                                "\u001b[32m2026-01-13 01:20:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0889392d-28b6-4335-8b89-a3e9cb5f251a (test) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 01:20:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0889392d-28b6-4335-8b89-a3e9cb5f251a (test) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Create test project] *****************************************************",
                                "",
                                "TASK [Create test domain] ******************************************************",
                                "Tuesday 13 January 2026  01:20:22 +0000 (0:00:00.051)       0:00:00.051 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test-admin user] **************************************************",
                                "Tuesday 13 January 2026  01:20:24 +0000 (0:00:02.834)       0:00:02.885 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add manager role to user test-admin] *************************************",
                                "Tuesday 13 January 2026  01:20:28 +0000 (0:00:03.758)       0:00:06.644 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test project] *****************************************************",
                                "Tuesday 13 January 2026  01:20:35 +0000 (0:00:06.479)       0:00:13.123 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test user] ********************************************************",
                                "Tuesday 13 January 2026  01:20:39 +0000 (0:00:03.976)       0:00:17.100 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add member roles to user test] *******************************************",
                                "Tuesday 13 January 2026  01:20:43 +0000 (0:00:03.934)       0:00:21.034 ******* ",
                                "\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m",
                                "",
                                "TASK [Create test server group] ************************************************",
                                "Tuesday 13 January 2026  01:20:54 +0000 (0:00:11.122)       0:00:32.156 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create ssh security group] ***********************************************",
                                "Tuesday 13 January 2026  01:20:58 +0000 (0:00:04.112)       0:00:36.269 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to ssh security group] ******************************************",
                                "Tuesday 13 January 2026  01:21:03 +0000 (0:00:04.889)       0:00:41.159 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create icmp security group] **********************************************",
                                "Tuesday 13 January 2026  01:21:07 +0000 (0:00:04.071)       0:00:45.231 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to icmp security group] *****************************************",
                                "Tuesday 13 January 2026  01:21:11 +0000 (0:00:03.969)       0:00:49.200 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test keypair] *****************************************************",
                                "Tuesday 13 January 2026  01:21:15 +0000 (0:00:04.176)       0:00:53.377 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test network] *****************************************************",
                                "Tuesday 13 January 2026  01:21:19 +0000 (0:00:03.732)       0:00:57.110 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test subnet] ******************************************************",
                                "Tuesday 13 January 2026  01:21:23 +0000 (0:00:04.466)       0:01:01.577 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test router] ******************************************************",
                                "Tuesday 13 January 2026  01:21:29 +0000 (0:00:05.592)       0:01:07.169 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test instances] ***************************************************",
                                "Tuesday 13 January 2026  01:21:41 +0000 (0:00:11.733)       0:01:18.902 ******* ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "",
                                "STILL ALIVE [task 'Create test instances' is running] **************************",
                                "",
                                "STILL ALIVE [task 'Create test instances' is running] **************************",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "",
                                "STILL ALIVE [task 'Create test instances' is running] **************************",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "",
                                "STILL ALIVE [task 'Create test instances' is running] **************************",
                                "",
                                "STILL ALIVE [task 'Create test instances' is running] **************************",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Add metadata to instances] ***********************************************",
                                "Tuesday 13 January 2026  01:25:55 +0000 (0:04:14.951)       0:05:33.854 ******* ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Add tag to instances] ****************************************************",
                                "Tuesday 13 January 2026  01:26:18 +0000 (0:00:22.113)       0:05:55.968 ******* ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Create test volume] ******************************************************",
                                "Tuesday 13 January 2026  01:26:51 +0000 (0:00:33.491)       0:06:29.460 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Attach test volume] ******************************************************",
                                "Tuesday 13 January 2026  01:26:58 +0000 (0:00:06.858)       0:06:36.318 ******* ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create floating ip address] **********************************************",
                                "Tuesday 13 January 2026  01:27:11 +0000 (0:00:13.010)       0:06:49.328 ******* ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Print floating ip address] ***********************************************",
                                "Tuesday 13 January 2026  01:27:17 +0000 (0:00:05.645)       0:06:54.974 ******* ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"192.168.112.133\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:27:17 +0000 (0:00:00.036)       0:06:55.010 ******* ",
                                "=============================================================================== ",
                                "Create test instances ------------------------------------------------- 254.95s",
                                "Add tag to instances --------------------------------------------------- 33.49s",
                                "Add metadata to instances ---------------------------------------------- 22.11s",
                                "Attach test volume ----------------------------------------------------- 13.01s",
                                "Create test router ----------------------------------------------------- 11.73s",
                                "Add member roles to user test ------------------------------------------ 11.12s",
                                "Create test volume ------------------------------------------------------ 6.86s",
                                "Add manager role to user test-admin ------------------------------------- 6.48s",
                                "Create floating ip address ---------------------------------------------- 5.65s",
                                "Create test subnet ------------------------------------------------------ 5.59s",
                                "Create ssh security group ----------------------------------------------- 4.89s",
                                "Create test network ----------------------------------------------------- 4.47s",
                                "Add rule to icmp security group ----------------------------------------- 4.18s",
                                "Create test server group ------------------------------------------------ 4.11s",
                                "Add rule to ssh security group ------------------------------------------ 4.07s",
                                "Create test project ----------------------------------------------------- 3.98s",
                                "Create icmp security group ---------------------------------------------- 3.97s",
                                "Create test user -------------------------------------------------------- 3.93s",
                                "Create test-admin user -------------------------------------------------- 3.76s",
                                "Create test keypair ----------------------------------------------------- 3.73s",
                                "+ server_list",
                                "+ openstack --os-cloud test server list",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE | test=192.168.112.182, 192.168.200.225 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE | test=192.168.112.189, 192.168.200.82  | N/A (booted from volume) | SCS-1L-1 |",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE | test=192.168.112.197, 192.168.200.207 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE | test=192.168.112.184, 192.168.200.155 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE | test=192.168.112.133, 192.168.200.13  | N/A (booted from volume) | SCS-1L-1 |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "+ openstack --os-cloud test server show test",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-01-13T01:22:25.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.133, 192.168.200.13                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-01-13T01:21:49Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | b7c58971654cc5a37db85f5dfcac31fb8cab3c23a991b40125a50d5c                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 105b5d47-9d05-45ab-8759-11299bd19793                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-01-13T01:26:00Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='9a1dd45e-c084-46c9-9b33-74ce3d98308a'                                                                                                                                                                                                                                                                                                             |",
                                "|                                     | delete_on_termination='False', id='6b4757db-f21e-47fa-a113-8d503701f961'                                                                                                                                                                                                                                                                                                            |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-1",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-01-13T01:23:22.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.184, 192.168.200.155                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-01-13T01:22:48Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 299cc453120c89f48d302e3526bfcb16d08ef82d89c546a2b6cd9172                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 6dc66e2a-091f-4945-a07c-d12a01a3a7a2                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-01-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='6f2b9777-0ed8-44e1-ab2e-af2a66da4310'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-2",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-01-13T01:24:18.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.197, 192.168.200.207                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-01-13T01:23:43Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 3b5320a87cce66fed3f5927f5afa231c08cad5c9a8f54dc3b7cadf91                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | e7fd2462-0b34-4225-997f-2488fa25b38e                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-01-13T01:26:08Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='7da2bc0c-af16-4147-95c4-154cd1481cc8'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-3",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-01-13T01:25:01.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.189, 192.168.200.82                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-01-13T01:24:36Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 299cc453120c89f48d302e3526bfcb16d08ef82d89c546a2b6cd9172                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 8ac30647-bb69-4d10-9174-753873ad881b                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-01-13T01:26:13Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='3fa8b44a-0e91-46fa-a4f5-3b822a805fe8'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-4",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-01-13T01:25:43.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.182, 192.168.200.225                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-01-13T01:25:18Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | b7c58971654cc5a37db85f5dfcac31fb8cab3c23a991b40125a50d5c                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 1b15012c-1cbf-4117-a3aa-ee944b91d5d2                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | 6947ea1887d64589bd716729c60f7645                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-01-13T01:26:17Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | cf41a0e0bd334268961c092230b5c237                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='484c345a-1c17-48f4-8619-af4ffc6b78e7'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.197",
                                "PING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=9.85 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.53 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.71 ms",
                                "",
                                "--- 192.168.112.197 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.714/4.698/9.853/3.660 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=7.24 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=3.01 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.33 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.334/3.863/7.243/2.486 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.66 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.89 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.30 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.298/2.947/5.658/1.931 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.189",
                                "PING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=5.29 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.75 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.53 ms",
                                "",
                                "--- 192.168.112.189 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.525/2.855/5.291/1.724 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.133",
                                "PING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.22 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.68 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.38 ms",
                                "",
                                "--- 192.168.112.133 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.375/2.424/4.216/1.273 ms",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4",
                                "\u001b[32m2026-01-13 01:28:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:28:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.197",
                                "PING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=7.49 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=1.61 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.34 ms",
                                "",
                                "--- 192.168.112.197 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.341/3.482/7.494/2.838 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=4.36 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=1.88 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.54 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.536/2.592/4.360/1.257 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.10 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.41 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.17 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.171/2.227/4.100/1.327 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.189",
                                "PING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=6.53 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.64 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.39 ms",
                                "",
                                "--- 192.168.112.189 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.391/3.188/6.533/2.367 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.133",
                                "PING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.03 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.37 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.24 ms",
                                "",
                                "--- 192.168.112.133 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.243/2.213/4.026/1.282 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5",
                                "\u001b[32m2026-01-13 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:29:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:30:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.197",
                                "PING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=5.99 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.77 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.85 ms",
                                "",
                                "--- 192.168.112.197 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.846/3.535/5.993/1.778 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.14 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=1.60 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.36 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.363/3.032/6.136/2.196 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=3.73 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.89 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.32 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.318/2.315/3.734/1.030 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.189",
                                "PING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=4.87 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.51 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.13 ms",
                                "",
                                "--- 192.168.112.189 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.134/2.505/4.871/1.679 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.133",
                                "PING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=3.11 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=2.25 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.70 ms",
                                "",
                                "--- 192.168.112.133 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.701/2.354/3.111/0.580 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3",
                                "\u001b[32m2026-01-13 01:31:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e7fd2462-0b34-4225-997f-2488fa25b38e\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:32:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:33:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.197",
                                "PING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=4.01 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.79 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=2.05 ms",
                                "",
                                "--- 192.168.112.197 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.053/2.952/4.012/0.807 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.93 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.04 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.46 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 2.043/3.812/6.933/2.213 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=9.59 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.55 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.24 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.244/4.460/9.590/3.666 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.189",
                                "PING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=3.65 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=1.39 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.19 ms",
                                "",
                                "--- 192.168.112.189 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.191/2.075/3.646/1.113 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.133",
                                "PING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=3.97 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.71 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.58 ms",
                                "",
                                "--- 192.168.112.133 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.579/2.418/3.965/1.095 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4",
                                "\u001b[32m2026-01-13 01:34:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 1b15012c-1cbf-4117-a3aa-ee944b91d5d2\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 8ac30647-bb69-4d10-9174-753873ad881b\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:34:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 8ac30647-bb69-4d10-9174-753873ad881b (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e7fd2462-0b34-4225-997f-2488fa25b38e\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e7fd2462-0b34-4225-997f-2488fa25b38e (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 6dc66e2a-091f-4945-a07c-d12a01a3a7a2\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-01-13 01:35:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 105b5d47-9d05-45ab-8759-11299bd19793\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-01-13 01:36:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 105b5d47-9d05-45ab-8759-11299bd19793 (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 1b15012c-1cbf-4117-a3aa-ee944b91d5d2 | test-4 | ACTIVE   |",
                                "| 8ac30647-bb69-4d10-9174-753873ad881b | test-3 | ACTIVE   |",
                                "| e7fd2462-0b34-4225-997f-2488fa25b38e | test-2 | ACTIVE   |",
                                "| 6dc66e2a-091f-4945-a07c-d12a01a3a7a2 | test-1 | ACTIVE   |",
                                "| 105b5d47-9d05-45ab-8759-11299bd19793 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.197",
                                "PING 192.168.112.197 (192.168.112.197) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.197: icmp_seq=1 ttl=63 time=8.48 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=2 ttl=63 time=2.13 ms",
                                "64 bytes from 192.168.112.197: icmp_seq=3 ttl=63 time=1.60 ms",
                                "",
                                "--- 192.168.112.197 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.596/4.068/8.483/3.129 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=4.66 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.29 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.52 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.522/2.825/4.659/1.334 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=3.96 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.01 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.74 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.738/2.567/3.956/0.988 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.189",
                                "PING 192.168.112.189 (192.168.112.189) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.189: icmp_seq=1 ttl=63 time=4.39 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=2 ttl=63 time=2.19 ms",
                                "64 bytes from 192.168.112.189: icmp_seq=3 ttl=63 time=1.76 ms",
                                "",
                                "--- 192.168.112.189 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.763/2.782/4.391/1.151 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.133",
                                "PING 192.168.112.133 (192.168.112.133) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.133: icmp_seq=1 ttl=63 time=4.42 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=2 ttl=63 time=1.78 ms",
                                "64 bytes from 192.168.112.133: icmp_seq=3 ttl=63 time=1.83 ms",
                                "",
                                "--- 192.168.112.133 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.776/2.674/4.415/1.230 ms"
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001e-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:36:54.092650Z",
                            "start": "2026-01-13T01:15:27.335362Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001e",
                        "name": "Run checks"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.193.234",
                                "/opt/configuration/scripts/check/302-openstack-with-tempest.sh"
                            ],
                            "delta": "0:03:43.199075",
                            "end": "2026-01-13 01:40:37.681628",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed05/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.193.234 /opt/configuration/scripts/check/302-openstack-with-tempest.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001f-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 01:36:54.482553",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# Tempest\n\n+ set -e\n+ echo\n+ echo '# Tempest'\n+ echo\n+ osism apply tempest --skip-tags run-tempest\n\u001b[32m2026-01-13 01:37:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a9596858-7eea-4afc-9fd3-2a6010a5a83a (tempest) was prepared for execution.\u001b[0m\n\u001b[32m2026-01-13 01:37:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a9596858-7eea-4afc-9fd3-2a6010a5a83a (tempest) has been started and output is visible here.\u001b[0m\n\nPLAY [Run tempest] *************************************************************\n\nTASK [osism.validations.tempest : Create tempest workdir] **********************\nTuesday 13 January 2026  01:37:11 +0000 (0:00:00.239)       0:00:00.239 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest wrapper script] *****************\nTuesday 13 January 2026  01:37:12 +0000 (0:00:00.710)       0:00:00.949 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Check for existing tempest initialisation] ***\nTuesday 13 January 2026  01:37:13 +0000 (0:00:01.216)       0:00:02.166 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Init tempest] ********************************\nTuesday 13 January 2026  01:37:13 +0000 (0:00:00.428)       0:00:02.594 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve image IDs] ***************************\nTuesday 13 January 2026  01:37:35 +0000 (0:00:21.649)       0:00:24.244 ******* \n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m\n\nTASK [osism.validations.tempest : Assert images have been resolved] ************\nTuesday 13 January 2026  01:37:44 +0000 (0:00:08.742)       0:00:32.987 ******* \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Get auth token] ******************************\nTuesday 13 January 2026  01:37:44 +0000 (0:00:00.156)       0:00:33.143 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get endpoint catalog] ************************\nTuesday 13 January 2026  01:37:48 +0000 (0:00:03.724)       0:00:36.867 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get service catalog] *************************\nTuesday 13 January 2026  01:37:49 +0000 (0:00:01.663)       0:00:38.531 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Register img_file name] **********************\nTuesday 13 January 2026  01:37:53 +0000 (0:00:03.622)       0:00:42.154 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Download img_file from image_ref] ************\nTuesday 13 January 2026  01:37:53 +0000 (0:00:00.189)       0:00:42.343 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Install qemu-utils package] ******************\nTuesday 13 January 2026  01:37:56 +0000 (0:00:02.579)       0:00:44.922 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Convert img_file to qcow2 format] ************\nTuesday 13 January 2026  01:38:06 +0000 (0:00:10.450)       0:00:55.373 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get network API extensions] ******************\nTuesday 13 January 2026  01:38:07 +0000 (0:00:00.767)       0:00:56.141 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Revoke token] ********************************\nTuesday 13 January 2026  01:38:09 +0000 (0:00:01.501)       0:00:57.642 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option api_extensions] ***\nTuesday 13 January 2026  01:38:10 +0000 (0:00:01.543)       0:00:59.186 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option img_file] *********\nTuesday 13 January 2026  01:38:10 +0000 (0:00:00.188)       0:00:59.375 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve floating network ID] *****************\nTuesday 13 January 2026  01:38:10 +0000 (0:00:00.180)       0:00:59.556 ******* \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Assert floating network id has been resolved] ***\nTuesday 13 January 2026  01:38:14 +0000 (0:00:03.761)       0:01:03.317 ******* \n\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Resolve flavor IDs] **************************\nTuesday 13 January 2026  01:38:14 +0000 (0:00:00.198)       0:01:03.515 ******* \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Assert flavors have been resolved] ***********\nTuesday 13 January 2026  01:38:15 +0000 (0:00:00.443)       0:01:03.958 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of exclude list] *******************\nTuesday 13 January 2026  01:38:15 +0000 (0:00:00.154)       0:01:04.113 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy exclude list] ***************************\nTuesday 13 January 2026  01:38:16 +0000 (0:00:00.531)       0:01:04.644 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of include list] *******************\nTuesday 13 January 2026  01:38:16 +0000 (0:00:00.905)       0:01:05.550 ******* \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy include list] ***************************\nTuesday 13 January 2026  01:38:17 +0000 (0:00:00.467)       0:01:06.017 ******* \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Create tempest flavors] **********************\nTuesday 13 January 2026  01:38:17 +0000 (0:00:00.158)       0:01:06.176 ******* \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest.conf file] **********************\nTuesday 13 January 2026  01:38:25 +0000 (0:00:07.795)       0:01:13.971 ******* \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 13 January 2026  01:38:26 +0000 (0:00:01.117)       0:01:15.089 ******* \n=============================================================================== \nosism.validations.tempest : Init tempest ------------------------------- 21.65s\nosism.validations.tempest : Install qemu-utils package ----------------- 10.45s\nosism.validations.tempest : Resolve image IDs --------------------------- 8.74s\nosism.validations.tempest : Create tempest flavors ---------------------- 7.80s\nosism.validations.tempest : Resolve floating network ID ----------------- 3.76s\nosism.validations.tempest : Get auth token ------------------------------ 3.72s\nosism.validations.tempest : Get service catalog ------------------------- 3.62s\nosism.validations.tempest : Download img_file from image_ref ------------ 2.58s\nosism.validations.tempest : Get endpoint catalog ------------------------ 1.66s\nosism.validations.tempest : Revoke token -------------------------------- 1.54s\nosism.validations.tempest : Get network API extensions ------------------ 1.50s\nosism.validations.tempest : Copy tempest wrapper script ----------------- 1.22s\nosism.validations.tempest : Copy tempest.conf file ---------------------- 1.12s\nosism.validations.tempest : Copy exclude list --------------------------- 0.91s\nosism.validations.tempest : Convert img_file to qcow2 format ------------ 0.77s\nosism.validations.tempest : Create tempest workdir ---------------------- 0.71s\nosism.validations.tempest : Get stats of exclude list ------------------- 0.53s\nosism.validations.tempest : Get stats of include list ------------------- 0.47s\nosism.validations.tempest : Resolve flavor IDs -------------------------- 0.44s\nosism.validations.tempest : Check for existing tempest initialisation --- 0.43s\n+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf\n+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf\n\n## IDENTITY (API)\n\n+ echo\n+ echo '## IDENTITY (API)'\n+ echo\n+ _tempest tempest.api.identity.v3\n+ local regex=tempest.api.identity.v3\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0138.log\n2026-01-13 01:38:31.936 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:38:32.039 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:38:32.039 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:38:35.858 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:38:35.866 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## IMAGE (API)\n\n+ echo\n+ echo '## IMAGE (API)'\n+ echo\n+ _tempest tempest.api.image.v2\n+ local regex=tempest.api.image.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0138.log\n2026-01-13 01:38:50.047 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:38:53.834 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:38:53.842 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## NETWORK (API)\n\n+ echo\n+ echo '## NETWORK (API)'\n+ echo\n+ _tempest tempest.api.network\n+ local regex=tempest.api.network\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0139.log\n2026-01-13 01:39:08.175 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:39:08.279 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:08.279 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:39:11.922 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:39:11.929 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## VOLUME (API)\n\n+ echo\n+ echo '## VOLUME (API)'\n+ echo\n+ _tempest tempest.api.volume\n+ local regex=tempest.api.volume\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0139.log\n2026-01-13 01:39:26.394 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:39:26.501 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:26.501 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:26.504 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:26.504 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:39:30.195 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:39:30.203 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## COMPUTE (API)\n\n+ echo\n+ echo '## COMPUTE (API)'\n+ echo\n+ _tempest tempest.api.compute\n+ local regex=tempest.api.compute\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0139.log\n2026-01-13 01:39:44.830 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:39:44.935 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:39:48.477 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:39:48.485 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:39:49.298 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:39:49.298 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## DNS (API)\n\n+ echo\n+ echo '## DNS (API)'\n+ echo\n+ _tempest designate_tempest_plugin.tests.api.v2\n+ local regex=designate_tempest_plugin.tests.api.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0139.log\n2026-01-13 01:40:02.714 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:40:02.812 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:40:02.815 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:40:02.815 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:40:06.536 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:40:06.544 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.\n\n## OBJECT-STORE (API)\n\n+ echo\n+ echo '## OBJECT-STORE (API)'\n+ echo\n+ _tempest tempest.api.object_storage\n+ local regex=tempest.api.object_storage\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260113-0140.log\n2026-01-13 01:40:22.189 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf\n2026-01-13 01:40:22.296 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:40:22.297 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:40:22.297 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n\n=========================\nFailures during discovery\n=========================\n--- stdout ---\n2026-01-13 01:40:25.957 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf\n2026-01-13 01:40:25.958 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests\n2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests\n2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests\n2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests\n2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate\n2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests\n2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests\n2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests\n2026-01-13 01:40:25.961 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests\n2026-01-13 01:40:25.961 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests\n2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests\n2026-01-13 01:40:25.965 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests\n2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests\n2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests\n2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests\n2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests\n--- import errors ---\nFailed to import test module: neutron_tempest_plugin.scenario.test_dns_integration\nTraceback (most recent call last):\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path\n    module = self._get_module_from_name(name)\n  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name\n    __import__(name)\n    ~~~~~~~~~~^^^^^^\n  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>\n    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')\n               ^^^^^^^^^^^^^^^^^^^^\nAttributeError: module 'testtools' has no attribute 'try_import'\n\n================================================================================\nThe above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                            "stdout_lines": [
                                "",
                                "# Tempest",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# Tempest'",
                                "+ echo",
                                "+ osism apply tempest --skip-tags run-tempest",
                                "\u001b[32m2026-01-13 01:37:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a9596858-7eea-4afc-9fd3-2a6010a5a83a (tempest) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-01-13 01:37:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a9596858-7eea-4afc-9fd3-2a6010a5a83a (tempest) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Run tempest] *************************************************************",
                                "",
                                "TASK [osism.validations.tempest : Create tempest workdir] **********************",
                                "Tuesday 13 January 2026  01:37:11 +0000 (0:00:00.239)       0:00:00.239 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest wrapper script] *****************",
                                "Tuesday 13 January 2026  01:37:12 +0000 (0:00:00.710)       0:00:00.949 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Check for existing tempest initialisation] ***",
                                "Tuesday 13 January 2026  01:37:13 +0000 (0:00:01.216)       0:00:02.166 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Init tempest] ********************************",
                                "Tuesday 13 January 2026  01:37:13 +0000 (0:00:00.428)       0:00:02.594 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve image IDs] ***************************",
                                "Tuesday 13 January 2026  01:37:35 +0000 (0:00:21.649)       0:00:24.244 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert images have been resolved] ************",
                                "Tuesday 13 January 2026  01:37:44 +0000 (0:00:08.742)       0:00:32.987 ******* ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get auth token] ******************************",
                                "Tuesday 13 January 2026  01:37:44 +0000 (0:00:00.156)       0:00:33.143 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get endpoint catalog] ************************",
                                "Tuesday 13 January 2026  01:37:48 +0000 (0:00:03.724)       0:00:36.867 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get service catalog] *************************",
                                "Tuesday 13 January 2026  01:37:49 +0000 (0:00:01.663)       0:00:38.531 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Register img_file name] **********************",
                                "Tuesday 13 January 2026  01:37:53 +0000 (0:00:03.622)       0:00:42.154 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Download img_file from image_ref] ************",
                                "Tuesday 13 January 2026  01:37:53 +0000 (0:00:00.189)       0:00:42.343 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Install qemu-utils package] ******************",
                                "Tuesday 13 January 2026  01:37:56 +0000 (0:00:02.579)       0:00:44.922 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Convert img_file to qcow2 format] ************",
                                "Tuesday 13 January 2026  01:38:06 +0000 (0:00:10.450)       0:00:55.373 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get network API extensions] ******************",
                                "Tuesday 13 January 2026  01:38:07 +0000 (0:00:00.767)       0:00:56.141 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Revoke token] ********************************",
                                "Tuesday 13 January 2026  01:38:09 +0000 (0:00:01.501)       0:00:57.642 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option api_extensions] ***",
                                "Tuesday 13 January 2026  01:38:10 +0000 (0:00:01.543)       0:00:59.186 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option img_file] *********",
                                "Tuesday 13 January 2026  01:38:10 +0000 (0:00:00.188)       0:00:59.375 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve floating network ID] *****************",
                                "Tuesday 13 January 2026  01:38:10 +0000 (0:00:00.180)       0:00:59.556 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert floating network id has been resolved] ***",
                                "Tuesday 13 January 2026  01:38:14 +0000 (0:00:03.761)       0:01:03.317 ******* ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve flavor IDs] **************************",
                                "Tuesday 13 January 2026  01:38:14 +0000 (0:00:00.198)       0:01:03.515 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert flavors have been resolved] ***********",
                                "Tuesday 13 January 2026  01:38:15 +0000 (0:00:00.443)       0:01:03.958 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of exclude list] *******************",
                                "Tuesday 13 January 2026  01:38:15 +0000 (0:00:00.154)       0:01:04.113 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy exclude list] ***************************",
                                "Tuesday 13 January 2026  01:38:16 +0000 (0:00:00.531)       0:01:04.644 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of include list] *******************",
                                "Tuesday 13 January 2026  01:38:16 +0000 (0:00:00.905)       0:01:05.550 ******* ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy include list] ***************************",
                                "Tuesday 13 January 2026  01:38:17 +0000 (0:00:00.467)       0:01:06.017 ******* ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Create tempest flavors] **********************",
                                "Tuesday 13 January 2026  01:38:17 +0000 (0:00:00.158)       0:01:06.176 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest.conf file] **********************",
                                "Tuesday 13 January 2026  01:38:25 +0000 (0:00:07.795)       0:01:13.971 ******* ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 13 January 2026  01:38:26 +0000 (0:00:01.117)       0:01:15.089 ******* ",
                                "=============================================================================== ",
                                "osism.validations.tempest : Init tempest ------------------------------- 21.65s",
                                "osism.validations.tempest : Install qemu-utils package ----------------- 10.45s",
                                "osism.validations.tempest : Resolve image IDs --------------------------- 8.74s",
                                "osism.validations.tempest : Create tempest flavors ---------------------- 7.80s",
                                "osism.validations.tempest : Resolve floating network ID ----------------- 3.76s",
                                "osism.validations.tempest : Get auth token ------------------------------ 3.72s",
                                "osism.validations.tempest : Get service catalog ------------------------- 3.62s",
                                "osism.validations.tempest : Download img_file from image_ref ------------ 2.58s",
                                "osism.validations.tempest : Get endpoint catalog ------------------------ 1.66s",
                                "osism.validations.tempest : Revoke token -------------------------------- 1.54s",
                                "osism.validations.tempest : Get network API extensions ------------------ 1.50s",
                                "osism.validations.tempest : Copy tempest wrapper script ----------------- 1.22s",
                                "osism.validations.tempest : Copy tempest.conf file ---------------------- 1.12s",
                                "osism.validations.tempest : Copy exclude list --------------------------- 0.91s",
                                "osism.validations.tempest : Convert img_file to qcow2 format ------------ 0.77s",
                                "osism.validations.tempest : Create tempest workdir ---------------------- 0.71s",
                                "osism.validations.tempest : Get stats of exclude list ------------------- 0.53s",
                                "osism.validations.tempest : Get stats of include list ------------------- 0.47s",
                                "osism.validations.tempest : Resolve flavor IDs -------------------------- 0.44s",
                                "osism.validations.tempest : Check for existing tempest initialisation --- 0.43s",
                                "+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf",
                                "+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf",
                                "",
                                "## IDENTITY (API)",
                                "",
                                "+ echo",
                                "+ echo '## IDENTITY (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.identity.v3",
                                "+ local regex=tempest.api.identity.v3",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0138.log",
                                "2026-01-13 01:38:31.936 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:38:32.039 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:32.039 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:32.040 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:32.041 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:38:32.042 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:32.043 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:38:35.858 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:35.860 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:35.861 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:35.862 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:35.863 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:35.866 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:36.714 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## IMAGE (API)",
                                "",
                                "+ echo",
                                "+ echo '## IMAGE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.image.v2",
                                "+ local regex=tempest.api.image.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0138.log",
                                "2026-01-13 01:38:50.047 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:50.132 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:38:50.133 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:50.134 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:50.135 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:38:53.834 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:53.836 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:53.837 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:53.838 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:53.839 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:38:53.842 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:38:54.648 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:38:54.649 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## NETWORK (API)",
                                "",
                                "+ echo",
                                "+ echo '## NETWORK (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.network",
                                "+ local regex=tempest.api.network",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0139.log",
                                "2026-01-13 01:39:08.175 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:39:08.279 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:08.279 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:08.280 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:08.281 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:08.282 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:39:11.922 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:11.924 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:11.925 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:11.926 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:11.927 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:11.929 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:12.741 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:12.742 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## VOLUME (API)",
                                "",
                                "+ echo",
                                "+ echo '## VOLUME (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.volume",
                                "+ local regex=tempest.api.volume",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0139.log",
                                "2026-01-13 01:39:26.394 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:39:26.501 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:26.501 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:26.502 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:26.503 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:26.504 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:26.504 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:26.505 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:26.506 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:39:30.195 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:30.197 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:30.198 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:30.199 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:30.200 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:30.203 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:39:31.084 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:31.085 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## COMPUTE (API)",
                                "",
                                "+ echo",
                                "+ echo '## COMPUTE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.compute",
                                "+ local regex=tempest.api.compute",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0139.log",
                                "2026-01-13 01:39:44.830 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:39:44.935 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:44.936 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:44.937 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:44.938 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:44.939 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:39:48.477 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:48.479 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:48.480 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:48.481 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:48.482 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:39:48.485 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:39:49.298 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:39:49.298 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:39:49.299 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## DNS (API)",
                                "",
                                "+ echo",
                                "+ echo '## DNS (API)'",
                                "+ echo",
                                "+ _tempest designate_tempest_plugin.tests.api.v2",
                                "+ local regex=designate_tempest_plugin.tests.api.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0139.log",
                                "2026-01-13 01:40:02.714 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:40:02.812 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:02.813 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:02.814 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:02.815 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:02.815 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:40:02.816 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:02.817 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:40:06.536 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:06.538 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:06.539 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:06.540 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:06.541 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:06.544 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:07.373 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:07.374 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path.",
                                "",
                                "## OBJECT-STORE (API)",
                                "",
                                "+ echo",
                                "+ echo '## OBJECT-STORE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.object_storage",
                                "+ local regex=tempest.api.object_storage",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260113-0140.log",
                                "2026-01-13 01:40:22.189 1 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf",
                                "2026-01-13 01:40:22.296 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:22.297 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:22.297 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:40:22.298 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:22.299 1 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:22.300 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:22.301 1 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "",
                                "=========================",
                                "Failures during discovery",
                                "=========================",
                                "--- stdout ---",
                                "2026-01-13 01:40:25.957 10 INFO tempest [-] Using tempest config file /tempest/etc/tempest.conf",
                                "2026-01-13 01:40:25.958 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:25.959 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: designate",
                                "2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:25.960 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:25.961 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:25.961 10 INFO tempest.test_discover.plugins [-] Register additional config options from Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: glance_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: designate",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: manila_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:25.962 10 INFO tempest.test_discover.plugins [-] List additional config options registered by Tempest plugin: ironic_tests",
                                "2026-01-13 01:40:25.965 10 WARNING oslo_config.cfg [-] Deprecated: Option \"auth_version\" from group \"identity\" is deprecated for removal (Identity v2 API was removed and v3 is the only available identity API version now).  Its value may be silently ignored in the future.",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: telemetry_tests",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: barbican_tests",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: glance_tests",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: octavia-tempest-plugin",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: magnum_tests",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: designate",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: neutron_tests",
                                "2026-01-13 01:40:26.800 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: cinder_tests",
                                "2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: manila_tests",
                                "2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: keystone_tests",
                                "2026-01-13 01:40:26.801 10 INFO tempest.test_discover.plugins [-] Loading tests from Tempest plugin: ironic_tests",
                                "--- import errors ---",
                                "Failed to import test module: neutron_tempest_plugin.scenario.test_dns_integration",
                                "Traceback (most recent call last):",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 396, in _find_test_path",
                                "    module = self._get_module_from_name(name)",
                                "  File \"/usr/local/lib/python3.13/unittest/loader.py\", line 339, in _get_module_from_name",
                                "    __import__(name)",
                                "    ~~~~~~~~~~^^^^^^",
                                "  File \"/usr/local/lib/python3.13/site-packages/neutron_tempest_plugin/scenario/test_dns_integration.py\", line 40, in <module>",
                                "    dns_base = testtools.try_import('designate_tempest_plugin.tests.base')",
                                "               ^^^^^^^^^^^^^^^^^^^^",
                                "AttributeError: module 'testtools' has no attribute 'try_import'",
                                "",
                                "================================================================================",
                                "The above traceback was encountered during test discovery which imports all the found test modules in the specified test_path."
                            ],
                            "zuul_log_id": "0242ac1f-6606-8e06-7a3c-00000000001f-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:37.814166Z",
                            "start": "2026-01-13T01:36:54.143972Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-00000000001f",
                        "name": "Run tempest"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_prometheus_alert_status | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:38.375052Z",
                            "start": "2026-01-13T01:40:37.865184Z"
                        },
                        "id": "0242ac1f-6606-8e06-7a3c-000000000020",
                        "name": "Check prometheus alert status"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 12,
            "failures": 0,
            "ignored": 0,
            "ok": 25,
            "rescued": 0,
            "skipped": 4,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:40:42.229382Z",
                    "start": "2026-01-13T01:40:39.468349Z"
                },
                "id": "0242ac1f-6606-92bd-c374-000000000002",
                "name": "Post output play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:39.540469Z",
                            "start": "2026-01-13T01:40:39.485640Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000007",
                        "name": "Register sources"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "sudo",
                                "-n",
                                "true"
                            ],
                            "delta": "0:00:00.013619",
                            "end": "2026-01-13 01:40:40.515665",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "sudo -n true",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-92bd-c374-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-01-13 01:40:40.502046",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sudo: a password is required",
                            "stdout_lines": [
                                "sudo: a password is required"
                            ],
                            "zuul_log_id": "0242ac1f-6606-92bd-c374-000000000009-1-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:40.586730Z",
                            "start": "2026-01-13T01:40:39.576376Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000009",
                        "name": "Check sudo"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:40.617889Z",
                            "start": "2026-01-13T01:40:40.593574Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000000a",
                        "name": "Set source and destination for files and folders"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "all_sources": []
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:40.703260Z",
                            "start": "2026-01-13T01:40:40.625443Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000000c",
                        "name": "Build a list of source, dest dictionaries"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed05/zuul-output/docs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed05/zuul-output/docs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1006,
                                    "group": "zuul-testbed05",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed05",
                                            "path": "/home/zuul-testbed05/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed05",
                                    "path": "/home/zuul-testbed05/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1005,
                                    "zj_output_dirs": "docs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed05/zuul-output/artifacts"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed05/zuul-output/artifacts"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1006,
                                    "group": "zuul-testbed05",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed05",
                                            "path": "/home/zuul-testbed05/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed05",
                                    "path": "/home/zuul-testbed05/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1005,
                                    "zj_output_dirs": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed05/zuul-output/logs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed05/zuul-output/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1006,
                                    "group": "zuul-testbed05",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed05",
                                            "path": "/home/zuul-testbed05/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed05",
                                    "path": "/home/zuul-testbed05/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1005,
                                    "zj_output_dirs": "logs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:41.745390Z",
                            "start": "2026-01-13T01:40:40.714024Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000000d",
                        "name": "Ensure target folders exist"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "results": [],
                            "skipped": true,
                            "skipped_reason": "No items in the list"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:41.776402Z",
                            "start": "2026-01-13T01:40:41.751644Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000000f",
                        "name": "Copy files and folders to staging folder"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed05/zuul-output/logs"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed05/zuul-output/logs"
                                }
                            },
                            "gid": 1006,
                            "group": "zuul-testbed05",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "dest": "/home/zuul-testbed05/zuul-output/logs",
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": "u=rwX,g=rX,o=rX",
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed05/zuul-output/logs",
                                    "recurse": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed05",
                            "path": "/home/zuul-testbed05/zuul-output/logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1005
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.099470Z",
                            "start": "2026-01-13T01:40:41.790590Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000011",
                        "name": "Make all log files readable"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stage_output_renames",
                            "changed": false,
                            "false_condition": "extensions_to_txt is defined and extensions_to_txt is not none",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.145891Z",
                            "start": "2026-01-13T01:40:42.110143Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000012",
                        "name": "Rename log files that match extensions_to_txt"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "find",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.190729Z",
                            "start": "2026-01-13T01:40:42.164910Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000014",
                        "name": "Discover log files for compression"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-92bd-c374-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.229382Z",
                            "start": "2026-01-13T01:40:42.203983Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000015",
                        "name": "Archive everything from logs"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:41:03.462271Z",
                    "start": "2026-01-13T01:40:42.243219Z"
                },
                "id": "0242ac1f-6606-92bd-c374-000000000017",
                "name": "Post cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.322085Z",
                            "start": "2026-01-13T01:40:42.254098Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-000000000019",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:40:42.378207Z",
                            "start": "2026-01-13T01:40:42.332831Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000001a",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:20.620708",
                            "end": "2026-01-13 01:41:03.322173",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-92bd-c374-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 01:40:42.701465",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-01-13 01:40:43 - clean up servers\n2026-01-13 01:40:44 - testbed-manager\n2026-01-13 01:40:44 - testbed-node-0\n2026-01-13 01:40:44 - testbed-node-2\n2026-01-13 01:40:44 - testbed-node-4\n2026-01-13 01:40:44 - testbed-node-3\n2026-01-13 01:40:44 - testbed-node-5\n2026-01-13 01:40:44 - testbed-node-1\n2026-01-13 01:40:44 - clean up keypairs\n2026-01-13 01:40:45 - testbed\n2026-01-13 01:40:45 - wait for servers to be gone\n2026-01-13 01:40:55 - clean up ports\n2026-01-13 01:40:56 - 033250be-3730-4e28-8bae-574aa81261e3\n2026-01-13 01:40:56 - 21f6ad4e-2daf-49cf-9f3a-83bd124ef2b6\n2026-01-13 01:40:57 - 5133aec3-bcc8-4613-a099-273522a85686\n2026-01-13 01:40:57 - 93e66259-6123-483f-83e0-5ca60df99fb6\n2026-01-13 01:40:57 - b46751e7-e6e1-4843-8018-ec0fe26ce24c\n2026-01-13 01:40:57 - dc661a00-4fe6-413a-9683-d66479df831d\n2026-01-13 01:40:58 - f9c66452-e535-44f2-abde-2653138d6638\n2026-01-13 01:40:58 - clean up volumes\n2026-01-13 01:40:58 - testbed-volume-3-node-base\n2026-01-13 01:40:58 - testbed-volume-2-node-base\n2026-01-13 01:40:58 - testbed-volume-manager-base\n2026-01-13 01:40:58 - testbed-volume-1-node-base\n2026-01-13 01:40:58 - testbed-volume-4-node-base\n2026-01-13 01:40:58 - testbed-volume-5-node-base\n2026-01-13 01:40:58 - testbed-volume-0-node-base\n2026-01-13 01:40:58 - testbed-volume-8-node-5\n2026-01-13 01:40:58 - testbed-volume-1-node-4\n2026-01-13 01:40:58 - testbed-volume-2-node-5\n2026-01-13 01:40:58 - testbed-volume-4-node-4\n2026-01-13 01:40:58 - testbed-volume-7-node-4\n2026-01-13 01:40:59 - testbed-volume-5-node-5\n2026-01-13 01:40:59 - testbed-volume-6-node-3\n2026-01-13 01:40:59 - testbed-volume-3-node-3\n2026-01-13 01:40:59 - testbed-volume-0-node-3\n2026-01-13 01:40:59 - disconnect routers\n2026-01-13 01:40:59 - testbed\n2026-01-13 01:41:00 - clean up subnets\n2026-01-13 01:41:00 - subnet-testbed-management\n2026-01-13 01:41:00 - clean up networks\n2026-01-13 01:41:00 - net-testbed-management\n2026-01-13 01:41:01 - clean up security groups\n2026-01-13 01:41:01 - testbed-node\n2026-01-13 01:41:01 - testbed-management\n2026-01-13 01:41:01 - clean up floating ips\n2026-01-13 01:41:01 - 81.163.193.234\n2026-01-13 01:41:01 - clean up routers\n2026-01-13 01:41:01 - testbed",
                            "stdout_lines": [
                                "2026-01-13 01:40:43 - clean up servers",
                                "2026-01-13 01:40:44 - testbed-manager",
                                "2026-01-13 01:40:44 - testbed-node-0",
                                "2026-01-13 01:40:44 - testbed-node-2",
                                "2026-01-13 01:40:44 - testbed-node-4",
                                "2026-01-13 01:40:44 - testbed-node-3",
                                "2026-01-13 01:40:44 - testbed-node-5",
                                "2026-01-13 01:40:44 - testbed-node-1",
                                "2026-01-13 01:40:44 - clean up keypairs",
                                "2026-01-13 01:40:45 - testbed",
                                "2026-01-13 01:40:45 - wait for servers to be gone",
                                "2026-01-13 01:40:55 - clean up ports",
                                "2026-01-13 01:40:56 - 033250be-3730-4e28-8bae-574aa81261e3",
                                "2026-01-13 01:40:56 - 21f6ad4e-2daf-49cf-9f3a-83bd124ef2b6",
                                "2026-01-13 01:40:57 - 5133aec3-bcc8-4613-a099-273522a85686",
                                "2026-01-13 01:40:57 - 93e66259-6123-483f-83e0-5ca60df99fb6",
                                "2026-01-13 01:40:57 - b46751e7-e6e1-4843-8018-ec0fe26ce24c",
                                "2026-01-13 01:40:57 - dc661a00-4fe6-413a-9683-d66479df831d",
                                "2026-01-13 01:40:58 - f9c66452-e535-44f2-abde-2653138d6638",
                                "2026-01-13 01:40:58 - clean up volumes",
                                "2026-01-13 01:40:58 - testbed-volume-3-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-2-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-manager-base",
                                "2026-01-13 01:40:58 - testbed-volume-1-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-4-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-5-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-0-node-base",
                                "2026-01-13 01:40:58 - testbed-volume-8-node-5",
                                "2026-01-13 01:40:58 - testbed-volume-1-node-4",
                                "2026-01-13 01:40:58 - testbed-volume-2-node-5",
                                "2026-01-13 01:40:58 - testbed-volume-4-node-4",
                                "2026-01-13 01:40:58 - testbed-volume-7-node-4",
                                "2026-01-13 01:40:59 - testbed-volume-5-node-5",
                                "2026-01-13 01:40:59 - testbed-volume-6-node-3",
                                "2026-01-13 01:40:59 - testbed-volume-3-node-3",
                                "2026-01-13 01:40:59 - testbed-volume-0-node-3",
                                "2026-01-13 01:40:59 - disconnect routers",
                                "2026-01-13 01:40:59 - testbed",
                                "2026-01-13 01:41:00 - clean up subnets",
                                "2026-01-13 01:41:00 - subnet-testbed-management",
                                "2026-01-13 01:41:00 - clean up networks",
                                "2026-01-13 01:41:00 - net-testbed-management",
                                "2026-01-13 01:41:01 - clean up security groups",
                                "2026-01-13 01:41:01 - testbed-node",
                                "2026-01-13 01:41:01 - testbed-management",
                                "2026-01-13 01:41:01 - clean up floating ips",
                                "2026-01-13 01:41:01 - 81.163.193.234",
                                "2026-01-13 01:41:01 - clean up routers",
                                "2026-01-13 01:41:01 - testbed"
                            ],
                            "zuul_log_id": "0242ac1f-6606-92bd-c374-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:03.462271Z",
                            "start": "2026-01-13T01:40:42.430264Z"
                        },
                        "id": "0242ac1f-6606-92bd-c374-00000000001b",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 7,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/cleanup.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:41:07.100491Z",
                    "start": "2026-01-13T01:41:04.392901Z"
                },
                "id": "0242ac1f-6606-f8a3-3366-000000000002",
                "name": "Cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:04.477026Z",
                            "start": "2026-01-13T01:41:04.410429Z"
                        },
                        "id": "0242ac1f-6606-f8a3-3366-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:04.524524Z",
                            "start": "2026-01-13T01:41:04.488322Z"
                        },
                        "id": "0242ac1f-6606-f8a3-3366-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.500762",
                            "end": "2026-01-13 01:41:06.975013",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed05/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-f8a3-3366-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-01-13 01:41:05.474251",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-01-13 01:41:05 - clean up servers\n2026-01-13 01:41:06 - clean up keypairs\n2026-01-13 01:41:06 - wait for servers to be gone\n2026-01-13 01:41:06 - clean up ports\n2026-01-13 01:41:06 - clean up volumes\n2026-01-13 01:41:06 - disconnect routers\n2026-01-13 01:41:06 - clean up subnets\n2026-01-13 01:41:06 - clean up networks\n2026-01-13 01:41:06 - clean up security groups\n2026-01-13 01:41:06 - clean up floating ips\n2026-01-13 01:41:06 - clean up routers",
                            "stdout_lines": [
                                "2026-01-13 01:41:05 - clean up servers",
                                "2026-01-13 01:41:06 - clean up keypairs",
                                "2026-01-13 01:41:06 - wait for servers to be gone",
                                "2026-01-13 01:41:06 - clean up ports",
                                "2026-01-13 01:41:06 - clean up volumes",
                                "2026-01-13 01:41:06 - disconnect routers",
                                "2026-01-13 01:41:06 - clean up subnets",
                                "2026-01-13 01:41:06 - clean up networks",
                                "2026-01-13 01:41:06 - clean up security groups",
                                "2026-01-13 01:41:06 - clean up floating ips",
                                "2026-01-13 01:41:06 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-f8a3-3366-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:07.100491Z",
                            "start": "2026-01-13T01:41:04.586404Z"
                        },
                        "id": "0242ac1f-6606-f8a3-3366-000000000006",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "2",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post-fetch.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:41:11.814203Z",
                    "start": "2026-01-13T01:41:08.020040Z"
                },
                "id": "0242ac1f-6606-12f2-bd82-000000000002",
                "name": "Base post-fetch"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12f2-bd82-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:08.094777Z",
                            "start": "2026-01-13T01:41:08.036821Z"
                        },
                        "id": "0242ac1f-6606-12f2-bd82-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12f2-bd82-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:08.170502Z",
                            "start": "2026-01-13T01:41:08.114246Z"
                        },
                        "id": "0242ac1f-6606-12f2-bd82-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12f2-bd82-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:09.305754Z",
                            "start": "2026-01-13T01:41:08.179499Z"
                        },
                        "id": "0242ac1f-6606-12f2-bd82-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/logs/ /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/artifacts/ /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/docs/ /var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed05@81.163.193.140:/home/zuul-testbed05/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12f2-bd82-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:11.756418Z",
                            "start": "2026-01-13T01:41:09.320494Z"
                        },
                        "id": "0242ac1f-6606-12f2-bd82-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "msg": "All items skipped",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "docs"
                                }
                            ],
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12f2-bd82-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:11.814203Z",
                            "start": "2026-01-13T01:41:11.768984Z"
                        },
                        "id": "0242ac1f-6606-12f2-bd82-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 3,
            "rescued": 0,
            "skipped": 2,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "3",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-01-13T01:41:13.797416Z",
                    "start": "2026-01-13T01:41:12.761824Z"
                },
                "id": "0242ac1f-6606-544c-aec7-000000000002",
                "name": "Base post"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCh/WKRmWvvHxojusJHTg1MCf5FkUfeP7ZQuitGz7rJpJE/3SzpC9mi7uddKyfgOEjV6y/xk9et5urJTN10nh4RxJuAesMu9VcoNOmMA/7/Z1VhWYKg2RcCQ/KTaCgd/MwbuwLywAxeV7JIWHfP65f0s9u0BzLbrET2e1T74XrnUA2qoMsa7u9QWJV8nZsQfiqummY3Az+hgPQ/+R7XXK9fam6CL2SidqB0v7dfwDnZZjLsoBfrGpvDKyEEm7DkSyhvqaasBPHSWTn6qe0UV6yXaMKTUnkrxKcBmpjgwR+o0ipHGVo0Ocnpm+aPm8Tq4rIMLLykDkzms6HS9rXvS8P54lawn2Ou2yWlTOjkNDJlJJyeu8pE1BxvMAq5CNw1Xt1VfmM1Lawbm1AKa7kL29LNyzvNHPZ68/kS8uL7+NKEjV88b7Y6yuX0AQR5mt6unQUntEGWZwLW2OlPoxHKQcMXr3yT/CDCaCX6K48Tzttncyl9yThpKlq2Wf8mIQbGBG8= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed05/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul-testbed05",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCh/WKRmWvvHxojusJHTg1MCf5FkUfeP7ZQuitGz7rJpJE/3SzpC9mi7uddKyfgOEjV6y/xk9et5urJTN10nh4RxJuAesMu9VcoNOmMA/7/Z1VhWYKg2RcCQ/KTaCgd/MwbuwLywAxeV7JIWHfP65f0s9u0BzLbrET2e1T74XrnUA2qoMsa7u9QWJV8nZsQfiqummY3Az+hgPQ/+R7XXK9fam6CL2SidqB0v7dfwDnZZjLsoBfrGpvDKyEEm7DkSyhvqaasBPHSWTn6qe0UV6yXaMKTUnkrxKcBmpjgwR+o0ipHGVo0Ocnpm+aPm8Tq4rIMLLykDkzms6HS9rXvS8P54lawn2Ou2yWlTOjkNDJlJJyeu8pE1BxvMAq5CNw1Xt1VfmM1Lawbm1AKa7kL29LNyzvNHPZ68/kS8uL7+NKEjV88b7Y6yuX0AQR5mt6unQUntEGWZwLW2OlPoxHKQcMXr3yT/CDCaCX6K48Tzttncyl9yThpKlq2Wf8mIQbGBG8= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed05/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul-testbed05",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-544c-aec7-000000000005",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/ca76e7f80cbb4cdda68907de4afef11c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-01-13T01:41:13.797416Z",
                            "start": "2026-01-13T01:41:12.778158Z"
                        },
                        "id": "0242ac1f-6606-544c-aec7-000000000007",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
